Feb 25 13:20:41 crc systemd[1]: Starting Kubernetes Kubelet... Feb 25 13:20:41 crc restorecon[4677]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:41 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 13:20:42 crc restorecon[4677]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 13:20:42 crc restorecon[4677]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 25 13:20:42 crc kubenswrapper[4815]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 25 13:20:42 crc kubenswrapper[4815]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 25 13:20:42 crc kubenswrapper[4815]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 25 13:20:42 crc kubenswrapper[4815]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 25 13:20:42 crc kubenswrapper[4815]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 25 13:20:42 crc kubenswrapper[4815]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.664334 4815 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670200 4815 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670239 4815 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670250 4815 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670263 4815 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670275 4815 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670284 4815 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670293 4815 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670307 4815 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670317 4815 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670326 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670335 4815 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670344 4815 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670352 4815 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670360 4815 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670368 4815 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670376 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670384 4815 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670392 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670399 4815 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670407 4815 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670414 4815 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670423 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670431 4815 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670439 4815 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670446 4815 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670454 4815 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670461 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670469 4815 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670477 4815 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670486 4815 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670494 4815 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670501 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670532 4815 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670540 4815 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670550 4815 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670561 4815 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670569 4815 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670577 4815 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670586 4815 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670595 4815 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670604 4815 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670611 4815 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670619 4815 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670627 4815 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670635 4815 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670642 4815 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670649 4815 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670657 4815 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670665 4815 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670672 4815 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670680 4815 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670688 4815 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670696 4815 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670704 4815 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670712 4815 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670719 4815 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670727 4815 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670736 4815 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670744 4815 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670752 4815 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670759 4815 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670766 4815 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670774 4815 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670781 4815 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670792 4815 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670801 4815 feature_gate.go:330] unrecognized feature gate: Example Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670809 4815 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670817 4815 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670828 4815 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670838 4815 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.670846 4815 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.671899 4815 flags.go:64] FLAG: --address="0.0.0.0" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.671922 4815 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.671938 4815 flags.go:64] FLAG: --anonymous-auth="true" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.671950 4815 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.671967 4815 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.671976 4815 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.671988 4815 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.671999 4815 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672009 4815 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672019 4815 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672028 4815 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672038 4815 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672047 4815 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672056 4815 flags.go:64] FLAG: --cgroup-root="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672065 4815 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672074 4815 flags.go:64] FLAG: --client-ca-file="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672082 4815 flags.go:64] FLAG: --cloud-config="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672091 4815 flags.go:64] FLAG: --cloud-provider="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672100 4815 flags.go:64] FLAG: --cluster-dns="[]" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672111 4815 flags.go:64] FLAG: --cluster-domain="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672120 4815 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672129 4815 flags.go:64] FLAG: --config-dir="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672139 4815 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672149 4815 flags.go:64] FLAG: --container-log-max-files="5" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672160 4815 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672169 4815 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672178 4815 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672187 4815 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672196 4815 flags.go:64] FLAG: --contention-profiling="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672204 4815 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672214 4815 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672223 4815 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672232 4815 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672243 4815 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672252 4815 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672261 4815 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672269 4815 flags.go:64] FLAG: --enable-load-reader="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672278 4815 flags.go:64] FLAG: --enable-server="true" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672287 4815 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672298 4815 flags.go:64] FLAG: --event-burst="100" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672308 4815 flags.go:64] FLAG: --event-qps="50" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672317 4815 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672326 4815 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672334 4815 flags.go:64] FLAG: --eviction-hard="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672345 4815 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672354 4815 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672363 4815 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672372 4815 flags.go:64] FLAG: --eviction-soft="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672381 4815 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672391 4815 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672400 4815 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672409 4815 flags.go:64] FLAG: --experimental-mounter-path="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672418 4815 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672427 4815 flags.go:64] FLAG: --fail-swap-on="true" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672435 4815 flags.go:64] FLAG: --feature-gates="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672446 4815 flags.go:64] FLAG: --file-check-frequency="20s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672455 4815 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672464 4815 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672474 4815 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672483 4815 flags.go:64] FLAG: --healthz-port="10248" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672492 4815 flags.go:64] FLAG: --help="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672501 4815 flags.go:64] FLAG: --hostname-override="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672536 4815 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672546 4815 flags.go:64] FLAG: --http-check-frequency="20s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672555 4815 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672564 4815 flags.go:64] FLAG: --image-credential-provider-config="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672572 4815 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672581 4815 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672590 4815 flags.go:64] FLAG: --image-service-endpoint="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672598 4815 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672607 4815 flags.go:64] FLAG: --kube-api-burst="100" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672616 4815 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672625 4815 flags.go:64] FLAG: --kube-api-qps="50" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672633 4815 flags.go:64] FLAG: --kube-reserved="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672643 4815 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672651 4815 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672660 4815 flags.go:64] FLAG: --kubelet-cgroups="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672668 4815 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672679 4815 flags.go:64] FLAG: --lock-file="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672687 4815 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672696 4815 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672705 4815 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672718 4815 flags.go:64] FLAG: --log-json-split-stream="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672727 4815 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672736 4815 flags.go:64] FLAG: --log-text-split-stream="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672745 4815 flags.go:64] FLAG: --logging-format="text" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672753 4815 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672763 4815 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672772 4815 flags.go:64] FLAG: --manifest-url="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672780 4815 flags.go:64] FLAG: --manifest-url-header="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672791 4815 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672800 4815 flags.go:64] FLAG: --max-open-files="1000000" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672811 4815 flags.go:64] FLAG: --max-pods="110" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672820 4815 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672830 4815 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672839 4815 flags.go:64] FLAG: --memory-manager-policy="None" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672847 4815 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672856 4815 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672865 4815 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672874 4815 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672893 4815 flags.go:64] FLAG: --node-status-max-images="50" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672902 4815 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672911 4815 flags.go:64] FLAG: --oom-score-adj="-999" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672921 4815 flags.go:64] FLAG: --pod-cidr="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672930 4815 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672943 4815 flags.go:64] FLAG: --pod-manifest-path="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672952 4815 flags.go:64] FLAG: --pod-max-pids="-1" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672963 4815 flags.go:64] FLAG: --pods-per-core="0" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672972 4815 flags.go:64] FLAG: --port="10250" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672981 4815 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672990 4815 flags.go:64] FLAG: --provider-id="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.672998 4815 flags.go:64] FLAG: --qos-reserved="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673007 4815 flags.go:64] FLAG: --read-only-port="10255" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673016 4815 flags.go:64] FLAG: --register-node="true" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673025 4815 flags.go:64] FLAG: --register-schedulable="true" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673034 4815 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673048 4815 flags.go:64] FLAG: --registry-burst="10" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673057 4815 flags.go:64] FLAG: --registry-qps="5" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673066 4815 flags.go:64] FLAG: --reserved-cpus="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673075 4815 flags.go:64] FLAG: --reserved-memory="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673086 4815 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673094 4815 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673103 4815 flags.go:64] FLAG: --rotate-certificates="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673112 4815 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673121 4815 flags.go:64] FLAG: --runonce="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673130 4815 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673139 4815 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673147 4815 flags.go:64] FLAG: --seccomp-default="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673157 4815 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673166 4815 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673175 4815 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673184 4815 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673193 4815 flags.go:64] FLAG: --storage-driver-password="root" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673202 4815 flags.go:64] FLAG: --storage-driver-secure="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673211 4815 flags.go:64] FLAG: --storage-driver-table="stats" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673220 4815 flags.go:64] FLAG: --storage-driver-user="root" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673228 4815 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673237 4815 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673246 4815 flags.go:64] FLAG: --system-cgroups="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673255 4815 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673268 4815 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673277 4815 flags.go:64] FLAG: --tls-cert-file="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673285 4815 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673296 4815 flags.go:64] FLAG: --tls-min-version="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673305 4815 flags.go:64] FLAG: --tls-private-key-file="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673313 4815 flags.go:64] FLAG: --topology-manager-policy="none" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673322 4815 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673331 4815 flags.go:64] FLAG: --topology-manager-scope="container" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673339 4815 flags.go:64] FLAG: --v="2" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673351 4815 flags.go:64] FLAG: --version="false" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673363 4815 flags.go:64] FLAG: --vmodule="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673373 4815 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.673382 4815 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673616 4815 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673629 4815 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673638 4815 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673647 4815 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673655 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673663 4815 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673670 4815 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673679 4815 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673686 4815 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673694 4815 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673701 4815 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673709 4815 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673717 4815 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673725 4815 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673733 4815 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673743 4815 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673753 4815 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673761 4815 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673768 4815 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673776 4815 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673785 4815 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673792 4815 feature_gate.go:330] unrecognized feature gate: Example Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673800 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673808 4815 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673815 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673832 4815 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673840 4815 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673848 4815 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673858 4815 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673868 4815 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673877 4815 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673886 4815 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673894 4815 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673903 4815 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673912 4815 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673922 4815 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673932 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673940 4815 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673949 4815 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673958 4815 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673965 4815 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673973 4815 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673981 4815 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673989 4815 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.673999 4815 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674009 4815 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674017 4815 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674025 4815 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674034 4815 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674043 4815 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674051 4815 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674059 4815 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674066 4815 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674074 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674082 4815 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674089 4815 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674096 4815 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674105 4815 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674113 4815 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674121 4815 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674128 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674135 4815 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674143 4815 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674153 4815 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674161 4815 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674169 4815 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674177 4815 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674185 4815 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674192 4815 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674200 4815 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.674207 4815 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.675200 4815 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.689448 4815 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.689539 4815 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689671 4815 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689694 4815 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689706 4815 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689719 4815 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689729 4815 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689740 4815 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689754 4815 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689763 4815 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689773 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689782 4815 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689790 4815 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689799 4815 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689807 4815 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689816 4815 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689825 4815 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689834 4815 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689845 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689855 4815 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689865 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689875 4815 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689885 4815 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689895 4815 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689905 4815 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689914 4815 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689923 4815 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689932 4815 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689941 4815 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689951 4815 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689959 4815 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689967 4815 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689975 4815 feature_gate.go:330] unrecognized feature gate: Example Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689983 4815 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689990 4815 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.689998 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690006 4815 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690013 4815 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690022 4815 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690030 4815 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690037 4815 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690046 4815 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690054 4815 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690062 4815 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690070 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690077 4815 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690086 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690093 4815 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690101 4815 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690110 4815 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690119 4815 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690128 4815 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690136 4815 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690144 4815 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690151 4815 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690159 4815 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690167 4815 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690178 4815 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690187 4815 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690195 4815 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690203 4815 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690212 4815 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690221 4815 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690232 4815 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690241 4815 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690250 4815 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690259 4815 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690268 4815 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690278 4815 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690286 4815 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690297 4815 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690306 4815 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690313 4815 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.690327 4815 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690626 4815 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690645 4815 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690655 4815 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690664 4815 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690672 4815 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690680 4815 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690687 4815 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690696 4815 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690704 4815 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690915 4815 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690925 4815 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690933 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690941 4815 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690949 4815 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690957 4815 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690966 4815 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690974 4815 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690982 4815 feature_gate.go:330] unrecognized feature gate: Example Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690990 4815 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.690998 4815 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691005 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691013 4815 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691021 4815 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691029 4815 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691036 4815 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691045 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691055 4815 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691063 4815 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691072 4815 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691081 4815 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691089 4815 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691096 4815 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691104 4815 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691112 4815 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691122 4815 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691133 4815 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691142 4815 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691152 4815 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691160 4815 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691168 4815 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691177 4815 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691185 4815 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691193 4815 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691201 4815 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691209 4815 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691217 4815 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691225 4815 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691232 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691240 4815 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691248 4815 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691256 4815 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691264 4815 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691275 4815 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691286 4815 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691296 4815 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691305 4815 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691313 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691322 4815 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691332 4815 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691342 4815 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691351 4815 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691359 4815 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691370 4815 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691379 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691411 4815 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691421 4815 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691429 4815 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691437 4815 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691445 4815 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691452 4815 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.691460 4815 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.691473 4815 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.693727 4815 server.go:940] "Client rotation is on, will bootstrap in background" Feb 25 13:20:42 crc kubenswrapper[4815]: E0225 13:20:42.701900 4815 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.706698 4815 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.706875 4815 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.708810 4815 server.go:997] "Starting client certificate rotation" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.708839 4815 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.709116 4815 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.735667 4815 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.738359 4815 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 25 13:20:42 crc kubenswrapper[4815]: E0225 13:20:42.738820 4815 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.142:6443: connect: connection refused" logger="UnhandledError" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.757337 4815 log.go:25] "Validated CRI v1 runtime API" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.797355 4815 log.go:25] "Validated CRI v1 image API" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.799747 4815 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.806449 4815 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-25-13-16-25-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.806524 4815 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.832037 4815 manager.go:217] Machine: {Timestamp:2026-02-25 13:20:42.826117287 +0000 UTC m=+0.627215431 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:86509d21-e518-46f8-a810-5a92ff458924 BootID:0c5638ab-7489-4aa7-9d40-f8fadb68ce90 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:91:17:e9 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:91:17:e9 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:e9:3c:4c Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:c6:30:53 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:f9:10:a4 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:d7:ea:7e Speed:-1 Mtu:1496} {Name:eth10 MacAddress:d2:25:b5:76:9e:21 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:02:d8:69:f1:1c:9a Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.832468 4815 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.832711 4815 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.833276 4815 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.833637 4815 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.833706 4815 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.834054 4815 topology_manager.go:138] "Creating topology manager with none policy" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.834074 4815 container_manager_linux.go:303] "Creating device plugin manager" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.834657 4815 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.834707 4815 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.836531 4815 state_mem.go:36] "Initialized new in-memory state store" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.836707 4815 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.840582 4815 kubelet.go:418] "Attempting to sync node with API server" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.840617 4815 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.840643 4815 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.840667 4815 kubelet.go:324] "Adding apiserver pod source" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.840686 4815 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.844882 4815 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.846929 4815 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.848553 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:20:42 crc kubenswrapper[4815]: E0225 13:20:42.848693 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.142:6443: connect: connection refused" logger="UnhandledError" Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.848733 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:20:42 crc kubenswrapper[4815]: E0225 13:20:42.848874 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.142:6443: connect: connection refused" logger="UnhandledError" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.849981 4815 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.852184 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.852226 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.852242 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.852255 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.852277 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.852292 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.852305 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.852326 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.852340 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.852356 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.852373 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.852386 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.853914 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.854538 4815 server.go:1280] "Started kubelet" Feb 25 13:20:42 crc systemd[1]: Started Kubernetes Kubelet. Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.857163 4815 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.857758 4815 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.857852 4815 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.857175 4815 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 25 13:20:42 crc kubenswrapper[4815]: E0225 13:20:42.858567 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.858738 4815 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.858766 4815 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.858924 4815 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.859206 4815 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.859425 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:20:42 crc kubenswrapper[4815]: E0225 13:20:42.860681 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" interval="200ms" Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.860746 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:20:42 crc kubenswrapper[4815]: E0225 13:20:42.861217 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.142:6443: connect: connection refused" logger="UnhandledError" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.862215 4815 server.go:460] "Adding debug handlers to kubelet server" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.868235 4815 factory.go:153] Registering CRI-O factory Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.868295 4815 factory.go:221] Registration of the crio container factory successfully Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.868576 4815 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.868609 4815 factory.go:55] Registering systemd factory Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.868629 4815 factory.go:221] Registration of the systemd container factory successfully Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.869303 4815 factory.go:103] Registering Raw factory Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.869352 4815 manager.go:1196] Started watching for new ooms in manager Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.878188 4815 manager.go:319] Starting recovery of all containers Feb 25 13:20:42 crc kubenswrapper[4815]: E0225 13:20:42.877989 4815 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.142:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.18977feeff396ef4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.854477556 +0000 UTC m=+0.655575640,LastTimestamp:2026-02-25 13:20:42.854477556 +0000 UTC m=+0.655575640,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885426 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885624 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885648 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885669 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885688 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885706 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885724 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885747 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885772 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885792 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885811 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885829 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885849 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885869 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885888 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885905 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885923 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885941 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885958 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885976 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.885993 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886010 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886035 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886053 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886070 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886135 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886171 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886200 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886225 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886251 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886277 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886299 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886316 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886337 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886354 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886372 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886395 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886417 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886435 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886454 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886471 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886489 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886533 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886552 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886572 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886590 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886610 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886628 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886647 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886666 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886684 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886702 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886729 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886749 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886768 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886789 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886807 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886826 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886847 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886866 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886885 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886902 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886924 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886942 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886960 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886977 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.886995 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887012 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887031 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887057 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887082 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887101 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887118 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887136 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887153 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887170 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887194 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887212 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887229 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887246 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887265 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887373 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887390 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887408 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887426 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887443 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887462 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887481 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887497 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887546 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887565 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887585 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887604 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887800 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887819 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887836 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887853 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887872 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887890 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887909 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887930 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887948 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887966 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.887985 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888012 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888034 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888055 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888077 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888109 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888130 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888150 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888173 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888195 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888213 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888232 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888251 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888269 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888286 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888306 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888325 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888342 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888360 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888379 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888397 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888416 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888438 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888457 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888473 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888492 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888611 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.888631 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891336 4815 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891420 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891454 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891481 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891501 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891556 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891578 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891604 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891628 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891648 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891670 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891688 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891741 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891760 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891780 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891799 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891839 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891860 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891880 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891901 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891918 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891937 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891959 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891978 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.891999 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892019 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892039 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892057 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892075 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892099 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892122 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892181 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892208 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892231 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892250 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892269 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892289 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892308 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892329 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892351 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892371 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892392 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892411 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892429 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892449 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892466 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892485 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892504 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892624 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892647 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892671 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892689 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892707 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892726 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892744 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892764 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892782 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892802 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892822 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892841 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892859 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892877 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892897 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892917 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892938 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892957 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892977 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.892996 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.893015 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.893033 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.893067 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.893087 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.893106 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.893128 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.893148 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.893167 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.893184 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.893212 4815 reconstruct.go:97] "Volume reconstruction finished" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.893227 4815 reconciler.go:26] "Reconciler: start to sync state" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.910166 4815 manager.go:324] Recovery completed Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.927787 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.928874 4815 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.930260 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.930338 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.930353 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.932130 4815 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.932156 4815 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.932183 4815 state_mem.go:36] "Initialized new in-memory state store" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.934201 4815 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.934260 4815 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.934302 4815 kubelet.go:2335] "Starting kubelet main sync loop" Feb 25 13:20:42 crc kubenswrapper[4815]: E0225 13:20:42.934757 4815 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 25 13:20:42 crc kubenswrapper[4815]: W0225 13:20:42.937838 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:20:42 crc kubenswrapper[4815]: E0225 13:20:42.937998 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.142:6443: connect: connection refused" logger="UnhandledError" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.950787 4815 policy_none.go:49] "None policy: Start" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.951961 4815 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 25 13:20:42 crc kubenswrapper[4815]: I0225 13:20:42.952007 4815 state_mem.go:35] "Initializing new in-memory state store" Feb 25 13:20:42 crc kubenswrapper[4815]: E0225 13:20:42.959471 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.013700 4815 manager.go:334] "Starting Device Plugin manager" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.013965 4815 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.013988 4815 server.go:79] "Starting device plugin registration server" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.014417 4815 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.014430 4815 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.014652 4815 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.014789 4815 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.014799 4815 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 25 13:20:43 crc kubenswrapper[4815]: E0225 13:20:43.026989 4815 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.035215 4815 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.035329 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.037118 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.037165 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.037181 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.037343 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.037731 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.037812 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.038284 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.038310 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.038322 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.038447 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.038599 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.038644 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.038930 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.038977 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.038997 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.039029 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.039047 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.039057 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.039215 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.039359 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.039402 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.039425 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.039472 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.039490 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.040035 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.040066 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.040080 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.040619 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.040770 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.040814 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.041389 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.041436 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.041454 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.041614 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.041645 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.041660 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.041706 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.041715 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.041740 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.041746 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.041751 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.042617 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.042645 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.042658 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:43 crc kubenswrapper[4815]: E0225 13:20:43.061743 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" interval="400ms" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.095132 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.095180 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.095241 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.095272 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.095296 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.095319 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.095340 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.095381 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.095485 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.095561 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.095604 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.095635 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.095669 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.095698 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.095727 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.114934 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.116691 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.116728 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.116741 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.116771 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 13:20:43 crc kubenswrapper[4815]: E0225 13:20:43.117313 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.142:6443: connect: connection refused" node="crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.196891 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.196983 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197040 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197088 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197129 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197164 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197204 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197227 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197169 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197161 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197239 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197336 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197462 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197544 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197595 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197684 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197717 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197792 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197818 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197918 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.197968 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.198001 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.198039 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.198062 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.198124 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.198187 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.198213 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.198262 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.198305 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.198121 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.318252 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.320099 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.320145 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.320165 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.320198 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 13:20:43 crc kubenswrapper[4815]: E0225 13:20:43.320753 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.142:6443: connect: connection refused" node="crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.383885 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.394174 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.415180 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: W0225 13:20:43.433166 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-8aafeb7712f085a937eb143b767635f6e6ccab7203c843fa85db2551a4057a26 WatchSource:0}: Error finding container 8aafeb7712f085a937eb143b767635f6e6ccab7203c843fa85db2551a4057a26: Status 404 returned error can't find the container with id 8aafeb7712f085a937eb143b767635f6e6ccab7203c843fa85db2551a4057a26 Feb 25 13:20:43 crc kubenswrapper[4815]: W0225 13:20:43.434378 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-9d5c858edc48ca8b3ae4241d7d90ff43455b2a0cc85381724c2116cd0b194ee2 WatchSource:0}: Error finding container 9d5c858edc48ca8b3ae4241d7d90ff43455b2a0cc85381724c2116cd0b194ee2: Status 404 returned error can't find the container with id 9d5c858edc48ca8b3ae4241d7d90ff43455b2a0cc85381724c2116cd0b194ee2 Feb 25 13:20:43 crc kubenswrapper[4815]: W0225 13:20:43.441959 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-184cf91092f91f61be3445523f0ba56b9c36b64ffb4cb2812a6c349f80d3587c WatchSource:0}: Error finding container 184cf91092f91f61be3445523f0ba56b9c36b64ffb4cb2812a6c349f80d3587c: Status 404 returned error can't find the container with id 184cf91092f91f61be3445523f0ba56b9c36b64ffb4cb2812a6c349f80d3587c Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.442498 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.451527 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:20:43 crc kubenswrapper[4815]: E0225 13:20:43.463475 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" interval="800ms" Feb 25 13:20:43 crc kubenswrapper[4815]: W0225 13:20:43.465438 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-336339a05daf7970ce20b183e93d77dd3bbb76e2d28a7bb10aa686b8233a2f46 WatchSource:0}: Error finding container 336339a05daf7970ce20b183e93d77dd3bbb76e2d28a7bb10aa686b8233a2f46: Status 404 returned error can't find the container with id 336339a05daf7970ce20b183e93d77dd3bbb76e2d28a7bb10aa686b8233a2f46 Feb 25 13:20:43 crc kubenswrapper[4815]: W0225 13:20:43.476010 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-1bdc538762fec7de4756e3da05d1925617f7b93f54d2552ee2216773d5917935 WatchSource:0}: Error finding container 1bdc538762fec7de4756e3da05d1925617f7b93f54d2552ee2216773d5917935: Status 404 returned error can't find the container with id 1bdc538762fec7de4756e3da05d1925617f7b93f54d2552ee2216773d5917935 Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.721307 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.723447 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.723544 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.723564 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.723610 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 13:20:43 crc kubenswrapper[4815]: E0225 13:20:43.724293 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.142:6443: connect: connection refused" node="crc" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.860249 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:20:43 crc kubenswrapper[4815]: W0225 13:20:43.877796 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:20:43 crc kubenswrapper[4815]: E0225 13:20:43.877922 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.142:6443: connect: connection refused" logger="UnhandledError" Feb 25 13:20:43 crc kubenswrapper[4815]: W0225 13:20:43.936469 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:20:43 crc kubenswrapper[4815]: E0225 13:20:43.936582 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.142:6443: connect: connection refused" logger="UnhandledError" Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.939539 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1bdc538762fec7de4756e3da05d1925617f7b93f54d2552ee2216773d5917935"} Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.940953 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"336339a05daf7970ce20b183e93d77dd3bbb76e2d28a7bb10aa686b8233a2f46"} Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.942277 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"184cf91092f91f61be3445523f0ba56b9c36b64ffb4cb2812a6c349f80d3587c"} Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.943551 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"9d5c858edc48ca8b3ae4241d7d90ff43455b2a0cc85381724c2116cd0b194ee2"} Feb 25 13:20:43 crc kubenswrapper[4815]: I0225 13:20:43.944732 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8aafeb7712f085a937eb143b767635f6e6ccab7203c843fa85db2551a4057a26"} Feb 25 13:20:44 crc kubenswrapper[4815]: W0225 13:20:44.039471 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:20:44 crc kubenswrapper[4815]: E0225 13:20:44.039647 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.142:6443: connect: connection refused" logger="UnhandledError" Feb 25 13:20:44 crc kubenswrapper[4815]: W0225 13:20:44.133875 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:20:44 crc kubenswrapper[4815]: E0225 13:20:44.134009 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.142:6443: connect: connection refused" logger="UnhandledError" Feb 25 13:20:44 crc kubenswrapper[4815]: E0225 13:20:44.265307 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" interval="1.6s" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.524447 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.525752 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.525790 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.525804 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.525831 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 13:20:44 crc kubenswrapper[4815]: E0225 13:20:44.526245 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.142:6443: connect: connection refused" node="crc" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.823651 4815 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 25 13:20:44 crc kubenswrapper[4815]: E0225 13:20:44.825043 4815 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.142:6443: connect: connection refused" logger="UnhandledError" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.861366 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.950270 4815 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="1a000fc74ffbe9a80c034cdb13a00d752e062e1aea3e3e5ff64f9774ffd0395a" exitCode=0 Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.950380 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"1a000fc74ffbe9a80c034cdb13a00d752e062e1aea3e3e5ff64f9774ffd0395a"} Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.950553 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.952070 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.952102 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.952114 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.954005 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b5e8323caee0f6b315fa6f46077cfe714a7498c4a4b9ada3b00bf5402705272c"} Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.954065 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e4a918c8db673c2f3de29af9a2f2d9265f880c8d4cc08f0f283804975bc9aa71"} Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.954078 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"90d91ca9797f7d8087787cf4ca1ad7951d125a330be7657f5b549f226021c931"} Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.954094 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5d0d690f29a96ea3e6b07d3e6e5da8f22972ff02b52434520a7da40b5d1fa4ee"} Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.954152 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.955671 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.955702 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.955716 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.956643 4815 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="aa9b163828c1d07b55e3911d56bb9432b4d8a6157b9b52a17b2d4264fc8e3b34" exitCode=0 Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.956683 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"aa9b163828c1d07b55e3911d56bb9432b4d8a6157b9b52a17b2d4264fc8e3b34"} Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.956874 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.957907 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.957929 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.957938 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.958849 4815 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="19311d205c069764f7d2d3ed470cf30dc0e47564b75ee0e51bc0fdcc5756965e" exitCode=0 Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.958939 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"19311d205c069764f7d2d3ed470cf30dc0e47564b75ee0e51bc0fdcc5756965e"} Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.959099 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.965625 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.965686 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.965671 4815 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="d2ae8fad65bcf342e9ead177f497797d3df16feb0cef1cd02a348db6258b261d" exitCode=0 Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.965624 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"d2ae8fad65bcf342e9ead177f497797d3df16feb0cef1cd02a348db6258b261d"} Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.965749 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.965710 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.967816 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.967870 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.967884 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.970012 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.971174 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.971208 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:44 crc kubenswrapper[4815]: I0225 13:20:44.971220 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:45 crc kubenswrapper[4815]: W0225 13:20:45.501089 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:20:45 crc kubenswrapper[4815]: E0225 13:20:45.501192 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.142:6443: connect: connection refused" logger="UnhandledError" Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.860837 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:20:45 crc kubenswrapper[4815]: E0225 13:20:45.866711 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" interval="3.2s" Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.969590 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f25f11a9ab520446a86fd17fe4c1ffc9bf0622c1bbf72019dbb1535caba77279"} Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.969632 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"38d1ca633d56103d9a2620de025cf43520585a65f1f225944a7a2f02714c8600"} Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.969643 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fa8220d4ac2c87c277e7e2a77c5d619bfb5af22a635106397a1c4104d420a70c"} Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.969652 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d4fb86989c8a3e44185969c0a59b6e39def5e539bf58cfd51a7f305ffb37ae26"} Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.970862 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"8eb5b6f86844d749ec2a7436ce7bf1cfa769a2daf71fa1939f0d4ce64e92cb99"} Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.970895 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.972404 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.972443 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.972456 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.974478 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"680c7818edadf07c8f5f188ddf1170a260a99953f09e6f9b5aced699991be064"} Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.974533 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3614e1a0f072bf6bc2de6783176401b83aa8dc404891e749161b450ce7350d53"} Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.974566 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7b8576c9bee9ce300cab6627700abe76083e395bb567907d231ecd77963c3e8b"} Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.974582 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.975321 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.975352 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.975365 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.976023 4815 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2e44d546ec9dc1d81b686a566da48d5a38d264a490447b73dbc8a5c6a308b978" exitCode=0 Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.976113 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2e44d546ec9dc1d81b686a566da48d5a38d264a490447b73dbc8a5c6a308b978"} Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.976143 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.976123 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.977084 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.977113 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.977124 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.977146 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.977163 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:45 crc kubenswrapper[4815]: I0225 13:20:45.977195 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.126681 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.127692 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.127717 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.127726 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.127747 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 13:20:46 crc kubenswrapper[4815]: E0225 13:20:46.128263 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.142:6443: connect: connection refused" node="crc" Feb 25 13:20:46 crc kubenswrapper[4815]: W0225 13:20:46.182966 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:20:46 crc kubenswrapper[4815]: E0225 13:20:46.183027 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.142:6443: connect: connection refused" logger="UnhandledError" Feb 25 13:20:46 crc kubenswrapper[4815]: E0225 13:20:46.198663 4815 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.142:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.18977feeff396ef4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.854477556 +0000 UTC m=+0.655575640,LastTimestamp:2026-02-25 13:20:42.854477556 +0000 UTC m=+0.655575640,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.984028 4815 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="6de963845af2a04ce3b19b5173d583e71c71dfef749a061d65406a0c4723a406" exitCode=0 Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.984197 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"6de963845af2a04ce3b19b5173d583e71c71dfef749a061d65406a0c4723a406"} Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.984205 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.985674 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.985709 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.985718 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.990167 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1e3f2d3cacb79f3fe0879f80a4bd35c7afdc997cc32906018313a5975b95937d"} Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.990234 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.990299 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.990832 4815 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.990903 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.991580 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.991617 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.991630 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.991591 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.991723 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.991746 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.992490 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.992558 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:46 crc kubenswrapper[4815]: I0225 13:20:46.992575 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:47 crc kubenswrapper[4815]: I0225 13:20:47.866366 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:20:47 crc kubenswrapper[4815]: I0225 13:20:47.866697 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:47 crc kubenswrapper[4815]: I0225 13:20:47.869930 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:47 crc kubenswrapper[4815]: I0225 13:20:47.870002 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:47 crc kubenswrapper[4815]: I0225 13:20:47.870023 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:47 crc kubenswrapper[4815]: I0225 13:20:47.997373 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"31666cef66ec933af08ff3649291929d4a4155389ad4d20330bc9f7161add489"} Feb 25 13:20:47 crc kubenswrapper[4815]: I0225 13:20:47.997428 4815 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 25 13:20:48 crc kubenswrapper[4815]: I0225 13:20:47.997459 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b9640cdc0698728157de1ee9aaf02f6d363abd53f675686af544b8db4601ec68"} Feb 25 13:20:48 crc kubenswrapper[4815]: I0225 13:20:47.997483 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:48 crc kubenswrapper[4815]: I0225 13:20:47.997490 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5a995b7f232eb72c84eb2505cfab176bf0924bbfef9940ef00b433aefca5fb37"} Feb 25 13:20:48 crc kubenswrapper[4815]: I0225 13:20:47.997547 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c47ba4a5cdd1c17b3002518f9c2c2dfd4b91dcc41409d7b5404aad23f6dacb51"} Feb 25 13:20:48 crc kubenswrapper[4815]: I0225 13:20:47.998474 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:48 crc kubenswrapper[4815]: I0225 13:20:47.998502 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:48 crc kubenswrapper[4815]: I0225 13:20:47.998529 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:49 crc kubenswrapper[4815]: I0225 13:20:49.006196 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"30d0312750a860fc4c7a6457ae7acb47bd217686cd56390ec0ad8f0f3f4983c1"} Feb 25 13:20:49 crc kubenswrapper[4815]: I0225 13:20:49.006325 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:49 crc kubenswrapper[4815]: I0225 13:20:49.007461 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:49 crc kubenswrapper[4815]: I0225 13:20:49.007680 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:49 crc kubenswrapper[4815]: I0225 13:20:49.007694 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:49 crc kubenswrapper[4815]: I0225 13:20:49.201853 4815 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 25 13:20:49 crc kubenswrapper[4815]: I0225 13:20:49.329247 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:49 crc kubenswrapper[4815]: I0225 13:20:49.331700 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:49 crc kubenswrapper[4815]: I0225 13:20:49.331748 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:49 crc kubenswrapper[4815]: I0225 13:20:49.331762 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:49 crc kubenswrapper[4815]: I0225 13:20:49.331795 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 13:20:49 crc kubenswrapper[4815]: I0225 13:20:49.625973 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:20:49 crc kubenswrapper[4815]: I0225 13:20:49.626179 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:49 crc kubenswrapper[4815]: I0225 13:20:49.627714 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:49 crc kubenswrapper[4815]: I0225 13:20:49.627743 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:49 crc kubenswrapper[4815]: I0225 13:20:49.627753 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:50 crc kubenswrapper[4815]: I0225 13:20:50.010323 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:50 crc kubenswrapper[4815]: I0225 13:20:50.011692 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:50 crc kubenswrapper[4815]: I0225 13:20:50.011748 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:50 crc kubenswrapper[4815]: I0225 13:20:50.011761 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:50 crc kubenswrapper[4815]: I0225 13:20:50.561283 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:20:50 crc kubenswrapper[4815]: I0225 13:20:50.561561 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:50 crc kubenswrapper[4815]: I0225 13:20:50.563086 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:50 crc kubenswrapper[4815]: I0225 13:20:50.563150 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:50 crc kubenswrapper[4815]: I0225 13:20:50.563169 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:50 crc kubenswrapper[4815]: I0225 13:20:50.677603 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:20:50 crc kubenswrapper[4815]: I0225 13:20:50.677811 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:50 crc kubenswrapper[4815]: I0225 13:20:50.679338 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:50 crc kubenswrapper[4815]: I0225 13:20:50.679391 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:50 crc kubenswrapper[4815]: I0225 13:20:50.679402 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:50 crc kubenswrapper[4815]: I0225 13:20:50.866654 4815 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 13:20:50 crc kubenswrapper[4815]: I0225 13:20:50.866760 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 13:20:50 crc kubenswrapper[4815]: I0225 13:20:50.994997 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:20:51 crc kubenswrapper[4815]: I0225 13:20:51.013456 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:51 crc kubenswrapper[4815]: I0225 13:20:51.014328 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:51 crc kubenswrapper[4815]: I0225 13:20:51.014359 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:51 crc kubenswrapper[4815]: I0225 13:20:51.014369 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:51 crc kubenswrapper[4815]: I0225 13:20:51.944882 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 25 13:20:51 crc kubenswrapper[4815]: I0225 13:20:51.945201 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:51 crc kubenswrapper[4815]: I0225 13:20:51.947044 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:51 crc kubenswrapper[4815]: I0225 13:20:51.947106 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:51 crc kubenswrapper[4815]: I0225 13:20:51.947119 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:52 crc kubenswrapper[4815]: I0225 13:20:52.371412 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:20:52 crc kubenswrapper[4815]: I0225 13:20:52.371629 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:52 crc kubenswrapper[4815]: I0225 13:20:52.372945 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:52 crc kubenswrapper[4815]: I0225 13:20:52.373027 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:52 crc kubenswrapper[4815]: I0225 13:20:52.373039 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:52 crc kubenswrapper[4815]: I0225 13:20:52.393312 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 13:20:52 crc kubenswrapper[4815]: I0225 13:20:52.393626 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:52 crc kubenswrapper[4815]: I0225 13:20:52.394973 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:52 crc kubenswrapper[4815]: I0225 13:20:52.395052 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:52 crc kubenswrapper[4815]: I0225 13:20:52.395080 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:53 crc kubenswrapper[4815]: E0225 13:20:53.027950 4815 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 13:20:53 crc kubenswrapper[4815]: I0225 13:20:53.298915 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:20:53 crc kubenswrapper[4815]: I0225 13:20:53.299145 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:53 crc kubenswrapper[4815]: I0225 13:20:53.300615 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:53 crc kubenswrapper[4815]: I0225 13:20:53.300669 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:53 crc kubenswrapper[4815]: I0225 13:20:53.300683 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:53 crc kubenswrapper[4815]: I0225 13:20:53.306018 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:20:53 crc kubenswrapper[4815]: I0225 13:20:53.560611 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 25 13:20:53 crc kubenswrapper[4815]: I0225 13:20:53.560842 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:53 crc kubenswrapper[4815]: I0225 13:20:53.562241 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:53 crc kubenswrapper[4815]: I0225 13:20:53.562300 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:53 crc kubenswrapper[4815]: I0225 13:20:53.562340 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:54 crc kubenswrapper[4815]: I0225 13:20:54.021160 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:54 crc kubenswrapper[4815]: I0225 13:20:54.022039 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:54 crc kubenswrapper[4815]: I0225 13:20:54.022083 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:54 crc kubenswrapper[4815]: I0225 13:20:54.022125 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:54 crc kubenswrapper[4815]: I0225 13:20:54.025786 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:20:55 crc kubenswrapper[4815]: I0225 13:20:55.023619 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:55 crc kubenswrapper[4815]: I0225 13:20:55.025611 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:55 crc kubenswrapper[4815]: I0225 13:20:55.025862 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:55 crc kubenswrapper[4815]: I0225 13:20:55.025998 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:56 crc kubenswrapper[4815]: W0225 13:20:56.639127 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:20:56Z is after 2026-02-23T05:33:13Z Feb 25 13:20:56 crc kubenswrapper[4815]: E0225 13:20:56.639205 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:20:56Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 13:20:56 crc kubenswrapper[4815]: W0225 13:20:56.640837 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:20:56Z is after 2026-02-23T05:33:13Z Feb 25 13:20:56 crc kubenswrapper[4815]: E0225 13:20:56.640948 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:20:56Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 13:20:56 crc kubenswrapper[4815]: W0225 13:20:56.642786 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:20:56Z is after 2026-02-23T05:33:13Z Feb 25 13:20:56 crc kubenswrapper[4815]: E0225 13:20:56.642834 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:20:56Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 13:20:56 crc kubenswrapper[4815]: I0225 13:20:56.644732 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:20:56Z is after 2026-02-23T05:33:13Z Feb 25 13:20:56 crc kubenswrapper[4815]: E0225 13:20:56.645059 4815 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:20:56Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.18977feeff396ef4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.854477556 +0000 UTC m=+0.655575640,LastTimestamp:2026-02-25 13:20:42.854477556 +0000 UTC m=+0.655575640,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:20:56 crc kubenswrapper[4815]: E0225 13:20:56.647496 4815 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:20:56Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 13:20:56 crc kubenswrapper[4815]: E0225 13:20:56.649465 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:20:56Z is after 2026-02-23T05:33:13Z" interval="6.4s" Feb 25 13:20:56 crc kubenswrapper[4815]: E0225 13:20:56.649983 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:20:56Z is after 2026-02-23T05:33:13Z" node="crc" Feb 25 13:20:56 crc kubenswrapper[4815]: I0225 13:20:56.650356 4815 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Feb 25 13:20:56 crc kubenswrapper[4815]: W0225 13:20:56.650432 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:20:56Z is after 2026-02-23T05:33:13Z Feb 25 13:20:56 crc kubenswrapper[4815]: I0225 13:20:56.650471 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 25 13:20:56 crc kubenswrapper[4815]: E0225 13:20:56.650493 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:20:56Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 13:20:56 crc kubenswrapper[4815]: I0225 13:20:56.657671 4815 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Feb 25 13:20:56 crc kubenswrapper[4815]: I0225 13:20:56.657730 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 25 13:20:56 crc kubenswrapper[4815]: I0225 13:20:56.865441 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:20:56Z is after 2026-02-23T05:33:13Z Feb 25 13:20:57 crc kubenswrapper[4815]: I0225 13:20:57.029597 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 25 13:20:57 crc kubenswrapper[4815]: I0225 13:20:57.031670 4815 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1e3f2d3cacb79f3fe0879f80a4bd35c7afdc997cc32906018313a5975b95937d" exitCode=255 Feb 25 13:20:57 crc kubenswrapper[4815]: I0225 13:20:57.031709 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1e3f2d3cacb79f3fe0879f80a4bd35c7afdc997cc32906018313a5975b95937d"} Feb 25 13:20:57 crc kubenswrapper[4815]: I0225 13:20:57.032065 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:57 crc kubenswrapper[4815]: I0225 13:20:57.033062 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:57 crc kubenswrapper[4815]: I0225 13:20:57.033113 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:57 crc kubenswrapper[4815]: I0225 13:20:57.033124 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:57 crc kubenswrapper[4815]: I0225 13:20:57.033756 4815 scope.go:117] "RemoveContainer" containerID="1e3f2d3cacb79f3fe0879f80a4bd35c7afdc997cc32906018313a5975b95937d" Feb 25 13:20:57 crc kubenswrapper[4815]: I0225 13:20:57.863429 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:20:57Z is after 2026-02-23T05:33:13Z Feb 25 13:20:58 crc kubenswrapper[4815]: I0225 13:20:58.036107 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 25 13:20:58 crc kubenswrapper[4815]: I0225 13:20:58.037847 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"72c5f9942cfc6be33df5ee01967ea703019a69de9a570e7346af40c8b7bc8605"} Feb 25 13:20:58 crc kubenswrapper[4815]: I0225 13:20:58.037999 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:58 crc kubenswrapper[4815]: I0225 13:20:58.038751 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:58 crc kubenswrapper[4815]: I0225 13:20:58.038779 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:58 crc kubenswrapper[4815]: I0225 13:20:58.038789 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:58 crc kubenswrapper[4815]: I0225 13:20:58.864475 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:20:58Z is after 2026-02-23T05:33:13Z Feb 25 13:20:59 crc kubenswrapper[4815]: I0225 13:20:59.048687 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 25 13:20:59 crc kubenswrapper[4815]: I0225 13:20:59.049362 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 25 13:20:59 crc kubenswrapper[4815]: I0225 13:20:59.051400 4815 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="72c5f9942cfc6be33df5ee01967ea703019a69de9a570e7346af40c8b7bc8605" exitCode=255 Feb 25 13:20:59 crc kubenswrapper[4815]: I0225 13:20:59.051448 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"72c5f9942cfc6be33df5ee01967ea703019a69de9a570e7346af40c8b7bc8605"} Feb 25 13:20:59 crc kubenswrapper[4815]: I0225 13:20:59.051490 4815 scope.go:117] "RemoveContainer" containerID="1e3f2d3cacb79f3fe0879f80a4bd35c7afdc997cc32906018313a5975b95937d" Feb 25 13:20:59 crc kubenswrapper[4815]: I0225 13:20:59.051613 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:20:59 crc kubenswrapper[4815]: I0225 13:20:59.052742 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:20:59 crc kubenswrapper[4815]: I0225 13:20:59.052790 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:20:59 crc kubenswrapper[4815]: I0225 13:20:59.052801 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:20:59 crc kubenswrapper[4815]: I0225 13:20:59.053616 4815 scope.go:117] "RemoveContainer" containerID="72c5f9942cfc6be33df5ee01967ea703019a69de9a570e7346af40c8b7bc8605" Feb 25 13:20:59 crc kubenswrapper[4815]: E0225 13:20:59.053860 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 13:20:59 crc kubenswrapper[4815]: I0225 13:20:59.626632 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:20:59 crc kubenswrapper[4815]: I0225 13:20:59.862788 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:20:59Z is after 2026-02-23T05:33:13Z Feb 25 13:21:00 crc kubenswrapper[4815]: I0225 13:21:00.055256 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 25 13:21:00 crc kubenswrapper[4815]: I0225 13:21:00.057964 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:00 crc kubenswrapper[4815]: I0225 13:21:00.058940 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:00 crc kubenswrapper[4815]: I0225 13:21:00.058969 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:00 crc kubenswrapper[4815]: I0225 13:21:00.058979 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:00 crc kubenswrapper[4815]: I0225 13:21:00.059490 4815 scope.go:117] "RemoveContainer" containerID="72c5f9942cfc6be33df5ee01967ea703019a69de9a570e7346af40c8b7bc8605" Feb 25 13:21:00 crc kubenswrapper[4815]: E0225 13:21:00.059683 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 13:21:00 crc kubenswrapper[4815]: I0225 13:21:00.687964 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:21:00 crc kubenswrapper[4815]: I0225 13:21:00.865112 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:00Z is after 2026-02-23T05:33:13Z Feb 25 13:21:00 crc kubenswrapper[4815]: I0225 13:21:00.867377 4815 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 13:21:00 crc kubenswrapper[4815]: I0225 13:21:00.867451 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 25 13:21:00 crc kubenswrapper[4815]: I0225 13:21:00.999268 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:21:01 crc kubenswrapper[4815]: I0225 13:21:01.061427 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:01 crc kubenswrapper[4815]: I0225 13:21:01.062980 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:01 crc kubenswrapper[4815]: I0225 13:21:01.063024 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:01 crc kubenswrapper[4815]: I0225 13:21:01.063057 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:01 crc kubenswrapper[4815]: I0225 13:21:01.063702 4815 scope.go:117] "RemoveContainer" containerID="72c5f9942cfc6be33df5ee01967ea703019a69de9a570e7346af40c8b7bc8605" Feb 25 13:21:01 crc kubenswrapper[4815]: E0225 13:21:01.063903 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 13:21:01 crc kubenswrapper[4815]: W0225 13:21:01.805749 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:01Z is after 2026-02-23T05:33:13Z Feb 25 13:21:01 crc kubenswrapper[4815]: E0225 13:21:01.805861 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:01Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 13:21:01 crc kubenswrapper[4815]: I0225 13:21:01.862484 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:01Z is after 2026-02-23T05:33:13Z Feb 25 13:21:02 crc kubenswrapper[4815]: I0225 13:21:02.063787 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:02 crc kubenswrapper[4815]: I0225 13:21:02.065160 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:02 crc kubenswrapper[4815]: I0225 13:21:02.065230 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:02 crc kubenswrapper[4815]: I0225 13:21:02.065255 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:02 crc kubenswrapper[4815]: I0225 13:21:02.066251 4815 scope.go:117] "RemoveContainer" containerID="72c5f9942cfc6be33df5ee01967ea703019a69de9a570e7346af40c8b7bc8605" Feb 25 13:21:02 crc kubenswrapper[4815]: E0225 13:21:02.066580 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 13:21:02 crc kubenswrapper[4815]: I0225 13:21:02.865312 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:02Z is after 2026-02-23T05:33:13Z Feb 25 13:21:02 crc kubenswrapper[4815]: W0225 13:21:02.933101 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:02Z is after 2026-02-23T05:33:13Z Feb 25 13:21:02 crc kubenswrapper[4815]: E0225 13:21:02.933190 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:02Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 13:21:03 crc kubenswrapper[4815]: E0225 13:21:03.028048 4815 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 13:21:03 crc kubenswrapper[4815]: I0225 13:21:03.050401 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:03 crc kubenswrapper[4815]: I0225 13:21:03.051609 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:03 crc kubenswrapper[4815]: I0225 13:21:03.051706 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:03 crc kubenswrapper[4815]: I0225 13:21:03.051727 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:03 crc kubenswrapper[4815]: I0225 13:21:03.051767 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 13:21:03 crc kubenswrapper[4815]: E0225 13:21:03.054035 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:03Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 25 13:21:03 crc kubenswrapper[4815]: E0225 13:21:03.055791 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:03Z is after 2026-02-23T05:33:13Z" node="crc" Feb 25 13:21:03 crc kubenswrapper[4815]: I0225 13:21:03.589754 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 25 13:21:03 crc kubenswrapper[4815]: I0225 13:21:03.591129 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:03 crc kubenswrapper[4815]: I0225 13:21:03.593110 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:03 crc kubenswrapper[4815]: I0225 13:21:03.593320 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:03 crc kubenswrapper[4815]: I0225 13:21:03.593461 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:03 crc kubenswrapper[4815]: I0225 13:21:03.605751 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 25 13:21:03 crc kubenswrapper[4815]: I0225 13:21:03.864863 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:03Z is after 2026-02-23T05:33:13Z Feb 25 13:21:04 crc kubenswrapper[4815]: I0225 13:21:04.069009 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:04 crc kubenswrapper[4815]: I0225 13:21:04.069824 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:04 crc kubenswrapper[4815]: I0225 13:21:04.069858 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:04 crc kubenswrapper[4815]: I0225 13:21:04.069870 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:04 crc kubenswrapper[4815]: I0225 13:21:04.803682 4815 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 25 13:21:04 crc kubenswrapper[4815]: E0225 13:21:04.807925 4815 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:04Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 13:21:04 crc kubenswrapper[4815]: I0225 13:21:04.864124 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:04Z is after 2026-02-23T05:33:13Z Feb 25 13:21:05 crc kubenswrapper[4815]: W0225 13:21:05.267008 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:05Z is after 2026-02-23T05:33:13Z Feb 25 13:21:05 crc kubenswrapper[4815]: E0225 13:21:05.267095 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:05Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 13:21:05 crc kubenswrapper[4815]: I0225 13:21:05.863334 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:05Z is after 2026-02-23T05:33:13Z Feb 25 13:21:06 crc kubenswrapper[4815]: E0225 13:21:06.648704 4815 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:06Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.18977feeff396ef4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.854477556 +0000 UTC m=+0.655575640,LastTimestamp:2026-02-25 13:20:42.854477556 +0000 UTC m=+0.655575640,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:06 crc kubenswrapper[4815]: I0225 13:21:06.863804 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:06Z is after 2026-02-23T05:33:13Z Feb 25 13:21:07 crc kubenswrapper[4815]: W0225 13:21:07.171558 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:07Z is after 2026-02-23T05:33:13Z Feb 25 13:21:07 crc kubenswrapper[4815]: E0225 13:21:07.171648 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:07Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 13:21:07 crc kubenswrapper[4815]: I0225 13:21:07.863458 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:07Z is after 2026-02-23T05:33:13Z Feb 25 13:21:08 crc kubenswrapper[4815]: I0225 13:21:08.864774 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:21:08 crc kubenswrapper[4815]: I0225 13:21:08.864860 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:08Z is after 2026-02-23T05:33:13Z Feb 25 13:21:08 crc kubenswrapper[4815]: I0225 13:21:08.865004 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:08 crc kubenswrapper[4815]: I0225 13:21:08.866392 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:08 crc kubenswrapper[4815]: I0225 13:21:08.866451 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:08 crc kubenswrapper[4815]: I0225 13:21:08.866465 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:08 crc kubenswrapper[4815]: I0225 13:21:08.867314 4815 scope.go:117] "RemoveContainer" containerID="72c5f9942cfc6be33df5ee01967ea703019a69de9a570e7346af40c8b7bc8605" Feb 25 13:21:09 crc kubenswrapper[4815]: I0225 13:21:09.087926 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 25 13:21:09 crc kubenswrapper[4815]: I0225 13:21:09.090779 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6656ddebba1eb92074357b7bfede1d0801bdb8773276da7f5e24f6d7d3a34408"} Feb 25 13:21:09 crc kubenswrapper[4815]: I0225 13:21:09.091107 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:09 crc kubenswrapper[4815]: I0225 13:21:09.092572 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:09 crc kubenswrapper[4815]: I0225 13:21:09.092611 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:09 crc kubenswrapper[4815]: I0225 13:21:09.092622 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:09 crc kubenswrapper[4815]: I0225 13:21:09.626314 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:21:09 crc kubenswrapper[4815]: W0225 13:21:09.657188 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:09Z is after 2026-02-23T05:33:13Z Feb 25 13:21:09 crc kubenswrapper[4815]: E0225 13:21:09.657317 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:09Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 13:21:09 crc kubenswrapper[4815]: I0225 13:21:09.862886 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:09Z is after 2026-02-23T05:33:13Z Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.055871 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.057331 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.057367 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.057378 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.057403 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 13:21:10 crc kubenswrapper[4815]: E0225 13:21:10.061636 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:10Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 25 13:21:10 crc kubenswrapper[4815]: E0225 13:21:10.065414 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:10Z is after 2026-02-23T05:33:13Z" node="crc" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.096406 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.096915 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.099231 4815 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6656ddebba1eb92074357b7bfede1d0801bdb8773276da7f5e24f6d7d3a34408" exitCode=255 Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.099276 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6656ddebba1eb92074357b7bfede1d0801bdb8773276da7f5e24f6d7d3a34408"} Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.099319 4815 scope.go:117] "RemoveContainer" containerID="72c5f9942cfc6be33df5ee01967ea703019a69de9a570e7346af40c8b7bc8605" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.099337 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.100094 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.100126 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.100136 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.100631 4815 scope.go:117] "RemoveContainer" containerID="6656ddebba1eb92074357b7bfede1d0801bdb8773276da7f5e24f6d7d3a34408" Feb 25 13:21:10 crc kubenswrapper[4815]: E0225 13:21:10.100803 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.863558 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:10Z is after 2026-02-23T05:33:13Z Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.867614 4815 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.867757 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.867831 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.868005 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.869421 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.869453 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.869468 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.869914 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"90d91ca9797f7d8087787cf4ca1ad7951d125a330be7657f5b549f226021c931"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Feb 25 13:21:10 crc kubenswrapper[4815]: I0225 13:21:10.870050 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://90d91ca9797f7d8087787cf4ca1ad7951d125a330be7657f5b549f226021c931" gracePeriod=30 Feb 25 13:21:10 crc kubenswrapper[4815]: W0225 13:21:10.993051 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:10Z is after 2026-02-23T05:33:13Z Feb 25 13:21:10 crc kubenswrapper[4815]: E0225 13:21:10.993191 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:10Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 13:21:11 crc kubenswrapper[4815]: I0225 13:21:11.103320 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 25 13:21:11 crc kubenswrapper[4815]: I0225 13:21:11.104937 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:11 crc kubenswrapper[4815]: I0225 13:21:11.106963 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 25 13:21:11 crc kubenswrapper[4815]: I0225 13:21:11.110722 4815 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="90d91ca9797f7d8087787cf4ca1ad7951d125a330be7657f5b549f226021c931" exitCode=255 Feb 25 13:21:11 crc kubenswrapper[4815]: I0225 13:21:11.110786 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"90d91ca9797f7d8087787cf4ca1ad7951d125a330be7657f5b549f226021c931"} Feb 25 13:21:11 crc kubenswrapper[4815]: I0225 13:21:11.110964 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:11 crc kubenswrapper[4815]: I0225 13:21:11.111001 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:11 crc kubenswrapper[4815]: I0225 13:21:11.111017 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:11 crc kubenswrapper[4815]: I0225 13:21:11.113566 4815 scope.go:117] "RemoveContainer" containerID="6656ddebba1eb92074357b7bfede1d0801bdb8773276da7f5e24f6d7d3a34408" Feb 25 13:21:11 crc kubenswrapper[4815]: E0225 13:21:11.113850 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 13:21:11 crc kubenswrapper[4815]: I0225 13:21:11.862935 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:11Z is after 2026-02-23T05:33:13Z Feb 25 13:21:12 crc kubenswrapper[4815]: I0225 13:21:12.119217 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 25 13:21:12 crc kubenswrapper[4815]: I0225 13:21:12.119900 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a2ea1e0385adb8a6e39d7d745efaab9b76e0ea04916d1fe5e13fab5dbd4818f2"} Feb 25 13:21:12 crc kubenswrapper[4815]: I0225 13:21:12.120074 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:12 crc kubenswrapper[4815]: I0225 13:21:12.121399 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:12 crc kubenswrapper[4815]: I0225 13:21:12.121448 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:12 crc kubenswrapper[4815]: I0225 13:21:12.121466 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:12 crc kubenswrapper[4815]: I0225 13:21:12.864640 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:12Z is after 2026-02-23T05:33:13Z Feb 25 13:21:13 crc kubenswrapper[4815]: E0225 13:21:13.029561 4815 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 13:21:13 crc kubenswrapper[4815]: I0225 13:21:13.122001 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:13 crc kubenswrapper[4815]: I0225 13:21:13.122996 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:13 crc kubenswrapper[4815]: I0225 13:21:13.123043 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:13 crc kubenswrapper[4815]: I0225 13:21:13.123064 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:13 crc kubenswrapper[4815]: I0225 13:21:13.863070 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:13Z is after 2026-02-23T05:33:13Z Feb 25 13:21:14 crc kubenswrapper[4815]: I0225 13:21:14.864942 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:14Z is after 2026-02-23T05:33:13Z Feb 25 13:21:15 crc kubenswrapper[4815]: I0225 13:21:15.863486 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:15Z is after 2026-02-23T05:33:13Z Feb 25 13:21:16 crc kubenswrapper[4815]: E0225 13:21:16.652358 4815 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:16Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.18977feeff396ef4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.854477556 +0000 UTC m=+0.655575640,LastTimestamp:2026-02-25 13:20:42.854477556 +0000 UTC m=+0.655575640,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:16 crc kubenswrapper[4815]: I0225 13:21:16.865034 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:16Z is after 2026-02-23T05:33:13Z Feb 25 13:21:17 crc kubenswrapper[4815]: I0225 13:21:17.065952 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:17 crc kubenswrapper[4815]: E0225 13:21:17.065998 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:17Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 25 13:21:17 crc kubenswrapper[4815]: I0225 13:21:17.068123 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:17 crc kubenswrapper[4815]: I0225 13:21:17.068153 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:17 crc kubenswrapper[4815]: I0225 13:21:17.068162 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:17 crc kubenswrapper[4815]: I0225 13:21:17.068184 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 13:21:17 crc kubenswrapper[4815]: E0225 13:21:17.072798 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:17Z is after 2026-02-23T05:33:13Z" node="crc" Feb 25 13:21:17 crc kubenswrapper[4815]: I0225 13:21:17.864976 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:17Z is after 2026-02-23T05:33:13Z Feb 25 13:21:17 crc kubenswrapper[4815]: I0225 13:21:17.867312 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:21:17 crc kubenswrapper[4815]: I0225 13:21:17.867501 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:17 crc kubenswrapper[4815]: I0225 13:21:17.869114 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:17 crc kubenswrapper[4815]: I0225 13:21:17.869163 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:17 crc kubenswrapper[4815]: I0225 13:21:17.869180 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:18 crc kubenswrapper[4815]: I0225 13:21:18.864674 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:21:18 crc kubenswrapper[4815]: I0225 13:21:18.864713 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:21:18Z is after 2026-02-23T05:33:13Z Feb 25 13:21:18 crc kubenswrapper[4815]: I0225 13:21:18.864886 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:18 crc kubenswrapper[4815]: I0225 13:21:18.866891 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:18 crc kubenswrapper[4815]: I0225 13:21:18.866927 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:18 crc kubenswrapper[4815]: I0225 13:21:18.866936 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:18 crc kubenswrapper[4815]: I0225 13:21:18.867460 4815 scope.go:117] "RemoveContainer" containerID="6656ddebba1eb92074357b7bfede1d0801bdb8773276da7f5e24f6d7d3a34408" Feb 25 13:21:18 crc kubenswrapper[4815]: E0225 13:21:18.867665 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 13:21:19 crc kubenswrapper[4815]: I0225 13:21:19.864373 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:20 crc kubenswrapper[4815]: I0225 13:21:20.562272 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:21:20 crc kubenswrapper[4815]: I0225 13:21:20.562610 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:20 crc kubenswrapper[4815]: I0225 13:21:20.564071 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:20 crc kubenswrapper[4815]: I0225 13:21:20.564120 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:20 crc kubenswrapper[4815]: I0225 13:21:20.564132 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:20 crc kubenswrapper[4815]: I0225 13:21:20.864988 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:20 crc kubenswrapper[4815]: I0225 13:21:20.868314 4815 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 13:21:20 crc kubenswrapper[4815]: I0225 13:21:20.868392 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 25 13:21:20 crc kubenswrapper[4815]: W0225 13:21:20.942953 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Feb 25 13:21:20 crc kubenswrapper[4815]: E0225 13:21:20.943018 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Feb 25 13:21:21 crc kubenswrapper[4815]: I0225 13:21:21.101402 4815 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 25 13:21:21 crc kubenswrapper[4815]: I0225 13:21:21.121139 4815 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 25 13:21:21 crc kubenswrapper[4815]: I0225 13:21:21.865156 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:22 crc kubenswrapper[4815]: I0225 13:21:22.868094 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:23 crc kubenswrapper[4815]: E0225 13:21:23.029708 4815 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 13:21:23 crc kubenswrapper[4815]: I0225 13:21:23.866696 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:24 crc kubenswrapper[4815]: I0225 13:21:24.072919 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:24 crc kubenswrapper[4815]: E0225 13:21:24.074258 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 25 13:21:24 crc kubenswrapper[4815]: I0225 13:21:24.075010 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:24 crc kubenswrapper[4815]: I0225 13:21:24.075084 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:24 crc kubenswrapper[4815]: I0225 13:21:24.075109 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:24 crc kubenswrapper[4815]: I0225 13:21:24.075147 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 13:21:24 crc kubenswrapper[4815]: E0225 13:21:24.081695 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 25 13:21:24 crc kubenswrapper[4815]: I0225 13:21:24.867975 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:25 crc kubenswrapper[4815]: I0225 13:21:25.866372 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.659992 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977feeff396ef4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.854477556 +0000 UTC m=+0.655575640,LastTimestamp:2026-02-25 13:20:42.854477556 +0000 UTC m=+0.655575640,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.665403 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bea776 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930317174 +0000 UTC m=+0.731415228,LastTimestamp:2026-02-25 13:20:42.930317174 +0000 UTC m=+0.731415228,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.671570 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bf1d5b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930347355 +0000 UTC m=+0.731445409,LastTimestamp:2026-02-25 13:20:42.930347355 +0000 UTC m=+0.731445409,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.677970 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bf5696 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930362006 +0000 UTC m=+0.731460060,LastTimestamp:2026-02-25 13:20:42.930362006 +0000 UTC m=+0.731460060,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.684159 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef08ede983 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:43.017300355 +0000 UTC m=+0.818398409,LastTimestamp:2026-02-25 13:20:43.017300355 +0000 UTC m=+0.818398409,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.687156 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bea776\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bea776 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930317174 +0000 UTC m=+0.731415228,LastTimestamp:2026-02-25 13:20:43.037147679 +0000 UTC m=+0.838245753,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.690858 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bf1d5b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bf1d5b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930347355 +0000 UTC m=+0.731445409,LastTimestamp:2026-02-25 13:20:43.03717423 +0000 UTC m=+0.838272294,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.693711 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bf5696\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bf5696 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930362006 +0000 UTC m=+0.731460060,LastTimestamp:2026-02-25 13:20:43.037189881 +0000 UTC m=+0.838287955,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.698358 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bea776\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bea776 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930317174 +0000 UTC m=+0.731415228,LastTimestamp:2026-02-25 13:20:43.038300117 +0000 UTC m=+0.839398171,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.702484 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bf1d5b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bf1d5b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930347355 +0000 UTC m=+0.731445409,LastTimestamp:2026-02-25 13:20:43.038318058 +0000 UTC m=+0.839416112,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.709254 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bf5696\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bf5696 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930362006 +0000 UTC m=+0.731460060,LastTimestamp:2026-02-25 13:20:43.038328618 +0000 UTC m=+0.839426672,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.713807 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bea776\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bea776 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930317174 +0000 UTC m=+0.731415228,LastTimestamp:2026-02-25 13:20:43.0389668 +0000 UTC m=+0.840064854,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.718761 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bf1d5b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bf1d5b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930347355 +0000 UTC m=+0.731445409,LastTimestamp:2026-02-25 13:20:43.038992691 +0000 UTC m=+0.840090745,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.725840 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bf5696\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bf5696 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930362006 +0000 UTC m=+0.731460060,LastTimestamp:2026-02-25 13:20:43.039002781 +0000 UTC m=+0.840100835,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.733726 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bea776\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bea776 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930317174 +0000 UTC m=+0.731415228,LastTimestamp:2026-02-25 13:20:43.039041112 +0000 UTC m=+0.840139166,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.739251 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bf1d5b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bf1d5b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930347355 +0000 UTC m=+0.731445409,LastTimestamp:2026-02-25 13:20:43.039053513 +0000 UTC m=+0.840151567,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.744025 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bf5696\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bf5696 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930362006 +0000 UTC m=+0.731460060,LastTimestamp:2026-02-25 13:20:43.039062313 +0000 UTC m=+0.840160367,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.748695 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bea776\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bea776 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930317174 +0000 UTC m=+0.731415228,LastTimestamp:2026-02-25 13:20:43.039450056 +0000 UTC m=+0.840548120,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.753832 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bf1d5b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bf1d5b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930347355 +0000 UTC m=+0.731445409,LastTimestamp:2026-02-25 13:20:43.039483917 +0000 UTC m=+0.840581991,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.757736 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bf5696\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bf5696 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930362006 +0000 UTC m=+0.731460060,LastTimestamp:2026-02-25 13:20:43.039498648 +0000 UTC m=+0.840596712,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.761335 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bea776\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bea776 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930317174 +0000 UTC m=+0.731415228,LastTimestamp:2026-02-25 13:20:43.040053117 +0000 UTC m=+0.841151171,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.766348 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bf1d5b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bf1d5b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930347355 +0000 UTC m=+0.731445409,LastTimestamp:2026-02-25 13:20:43.040073777 +0000 UTC m=+0.841171831,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.772097 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bf5696\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bf5696 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930362006 +0000 UTC m=+0.731460060,LastTimestamp:2026-02-25 13:20:43.040086538 +0000 UTC m=+0.841184592,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.777096 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bea776\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bea776 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930317174 +0000 UTC m=+0.731415228,LastTimestamp:2026-02-25 13:20:43.041417052 +0000 UTC m=+0.842515136,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.781975 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18977fef03bf1d5b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18977fef03bf1d5b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:42.930347355 +0000 UTC m=+0.731445409,LastTimestamp:2026-02-25 13:20:43.041446713 +0000 UTC m=+0.842544787,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.788493 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18977fef22294547 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:43.440620871 +0000 UTC m=+1.241718925,LastTimestamp:2026-02-25 13:20:43.440620871 +0000 UTC m=+1.241718925,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.794335 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18977fef222afbbb openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:43.440733115 +0000 UTC m=+1.241831209,LastTimestamp:2026-02-25 13:20:43.440733115 +0000 UTC m=+1.241831209,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.798640 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fef2280b45f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:43.446350943 +0000 UTC m=+1.247448997,LastTimestamp:2026-02-25 13:20:43.446350943 +0000 UTC m=+1.247448997,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.803582 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977fef23e295d4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:43.469542868 +0000 UTC m=+1.270640932,LastTimestamp:2026-02-25 13:20:43.469542868 +0000 UTC m=+1.270640932,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.809476 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977fef249a4295 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:43.481580181 +0000 UTC m=+1.282678285,LastTimestamp:2026-02-25 13:20:43.481580181 +0000 UTC m=+1.282678285,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.814002 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18977fef450d7bc2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.02600237 +0000 UTC m=+1.827100424,LastTimestamp:2026-02-25 13:20:44.02600237 +0000 UTC m=+1.827100424,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.819077 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977fef452bc740 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.027987776 +0000 UTC m=+1.829085830,LastTimestamp:2026-02-25 13:20:44.027987776 +0000 UTC m=+1.829085830,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.824632 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fef452e1e55 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.028141141 +0000 UTC m=+1.829239195,LastTimestamp:2026-02-25 13:20:44.028141141 +0000 UTC m=+1.829239195,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.830468 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18977fef45a7d0ee openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.036116718 +0000 UTC m=+1.837214772,LastTimestamp:2026-02-25 13:20:44.036116718 +0000 UTC m=+1.837214772,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.835123 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18977fef45f9809b openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.041470107 +0000 UTC m=+1.842568161,LastTimestamp:2026-02-25 13:20:44.041470107 +0000 UTC m=+1.842568161,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.839126 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977fef4616d253 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.043391571 +0000 UTC m=+1.844489625,LastTimestamp:2026-02-25 13:20:44.043391571 +0000 UTC m=+1.844489625,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.845069 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977fef4617d7b3 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.043458483 +0000 UTC m=+1.844556537,LastTimestamp:2026-02-25 13:20:44.043458483 +0000 UTC m=+1.844556537,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.849197 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fef46192756 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.043544406 +0000 UTC m=+1.844642460,LastTimestamp:2026-02-25 13:20:44.043544406 +0000 UTC m=+1.844642460,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.853037 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977fef462cdfac openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.04483678 +0000 UTC m=+1.845934834,LastTimestamp:2026-02-25 13:20:44.04483678 +0000 UTC m=+1.845934834,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.857290 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18977fef46c64f3a openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.054892346 +0000 UTC m=+1.855990400,LastTimestamp:2026-02-25 13:20:44.054892346 +0000 UTC m=+1.855990400,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.861948 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977fef47726b4c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.066171724 +0000 UTC m=+1.867269778,LastTimestamp:2026-02-25 13:20:44.066171724 +0000 UTC m=+1.867269778,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: I0225 13:21:26.862330 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.864538 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977fef565be9aa openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.316354986 +0000 UTC m=+2.117453040,LastTimestamp:2026-02-25 13:20:44.316354986 +0000 UTC m=+2.117453040,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.866470 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977fef56f01c82 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.32606733 +0000 UTC m=+2.127165384,LastTimestamp:2026-02-25 13:20:44.32606733 +0000 UTC m=+2.127165384,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.871735 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977fef570300ba openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.327305402 +0000 UTC m=+2.128403456,LastTimestamp:2026-02-25 13:20:44.327305402 +0000 UTC m=+2.128403456,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.873475 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977fef61392b75 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.498627445 +0000 UTC m=+2.299725509,LastTimestamp:2026-02-25 13:20:44.498627445 +0000 UTC m=+2.299725509,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.876678 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977fef61f2f2f5 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.510802677 +0000 UTC m=+2.311900751,LastTimestamp:2026-02-25 13:20:44.510802677 +0000 UTC m=+2.311900751,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.882098 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977fef620506d8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.511987416 +0000 UTC m=+2.313085470,LastTimestamp:2026-02-25 13:20:44.511987416 +0000 UTC m=+2.313085470,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.886257 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977fef6bb350fc openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.674404604 +0000 UTC m=+2.475502658,LastTimestamp:2026-02-25 13:20:44.674404604 +0000 UTC m=+2.475502658,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.892010 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977fef6c59c920 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.685314336 +0000 UTC m=+2.486412390,LastTimestamp:2026-02-25 13:20:44.685314336 +0000 UTC m=+2.486412390,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.895594 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18977fef7c569ae2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.953541346 +0000 UTC m=+2.754639400,LastTimestamp:2026-02-25 13:20:44.953541346 +0000 UTC m=+2.754639400,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.901632 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977fef7cb2cbcd openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.959583181 +0000 UTC m=+2.760681275,LastTimestamp:2026-02-25 13:20:44.959583181 +0000 UTC m=+2.760681275,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.906601 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fef7d4bf2c4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.969620164 +0000 UTC m=+2.770718208,LastTimestamp:2026-02-25 13:20:44.969620164 +0000 UTC m=+2.770718208,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.912598 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18977fef7d4d1aa3 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.969695907 +0000 UTC m=+2.770793961,LastTimestamp:2026-02-25 13:20:44.969695907 +0000 UTC m=+2.770793961,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.918254 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18977fef89c56e4b openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.178908235 +0000 UTC m=+2.980006289,LastTimestamp:2026-02-25 13:20:45.178908235 +0000 UTC m=+2.980006289,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.922776 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977fef89c8d088 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.179129992 +0000 UTC m=+2.980228046,LastTimestamp:2026-02-25 13:20:45.179129992 +0000 UTC m=+2.980228046,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.927658 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18977fef89ca64d8 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.179233496 +0000 UTC m=+2.980331560,LastTimestamp:2026-02-25 13:20:45.179233496 +0000 UTC m=+2.980331560,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.932476 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fef89cfe606 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.179594246 +0000 UTC m=+2.980692300,LastTimestamp:2026-02-25 13:20:45.179594246 +0000 UTC m=+2.980692300,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.936832 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18977fef8ae852ac openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.19797214 +0000 UTC m=+2.999070194,LastTimestamp:2026-02-25 13:20:45.19797214 +0000 UTC m=+2.999070194,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.940384 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18977fef8af8e621 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.199058465 +0000 UTC m=+3.000156519,LastTimestamp:2026-02-25 13:20:45.199058465 +0000 UTC m=+3.000156519,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.944096 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18977fef8b20f27e openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.20168307 +0000 UTC m=+3.002781114,LastTimestamp:2026-02-25 13:20:45.20168307 +0000 UTC m=+3.002781114,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.949914 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fef8b254e80 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.201968768 +0000 UTC m=+3.003066822,LastTimestamp:2026-02-25 13:20:45.201968768 +0000 UTC m=+3.003066822,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.954960 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977fef8b3ba192 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.203431826 +0000 UTC m=+3.004529880,LastTimestamp:2026-02-25 13:20:45.203431826 +0000 UTC m=+3.004529880,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.959921 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fef8b66582e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.206231086 +0000 UTC m=+3.007329140,LastTimestamp:2026-02-25 13:20:45.206231086 +0000 UTC m=+3.007329140,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.963915 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18977fef96e93be2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.399358434 +0000 UTC m=+3.200456488,LastTimestamp:2026-02-25 13:20:45.399358434 +0000 UTC m=+3.200456488,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.968305 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fef971732a2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.402370722 +0000 UTC m=+3.203468776,LastTimestamp:2026-02-25 13:20:45.402370722 +0000 UTC m=+3.203468776,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.973389 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18977fef97de8353 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.415433043 +0000 UTC m=+3.216531097,LastTimestamp:2026-02-25 13:20:45.415433043 +0000 UTC m=+3.216531097,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.977617 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18977fef97ef7c13 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.416545299 +0000 UTC m=+3.217643353,LastTimestamp:2026-02-25 13:20:45.416545299 +0000 UTC m=+3.217643353,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.981304 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fef980ab340 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.418328896 +0000 UTC m=+3.219426950,LastTimestamp:2026-02-25 13:20:45.418328896 +0000 UTC m=+3.219426950,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.985075 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fef981800a2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.419200674 +0000 UTC m=+3.220298728,LastTimestamp:2026-02-25 13:20:45.419200674 +0000 UTC m=+3.220298728,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.988875 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18977fefa27138a2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.592819874 +0000 UTC m=+3.393917928,LastTimestamp:2026-02-25 13:20:45.592819874 +0000 UTC m=+3.393917928,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.992555 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fefa2823139 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.593932089 +0000 UTC m=+3.395030153,LastTimestamp:2026-02-25 13:20:45.593932089 +0000 UTC m=+3.395030153,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:26 crc kubenswrapper[4815]: E0225 13:21:26.997019 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18977fefa36d3c3b openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.609335867 +0000 UTC m=+3.410433921,LastTimestamp:2026-02-25 13:20:45.609335867 +0000 UTC m=+3.410433921,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.001500 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fefa3dfac9a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.616835738 +0000 UTC m=+3.417933802,LastTimestamp:2026-02-25 13:20:45.616835738 +0000 UTC m=+3.417933802,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.005302 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fefa3f3aef5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.618147061 +0000 UTC m=+3.419245125,LastTimestamp:2026-02-25 13:20:45.618147061 +0000 UTC m=+3.419245125,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.009098 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fefaedafd18 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.80107804 +0000 UTC m=+3.602176084,LastTimestamp:2026-02-25 13:20:45.80107804 +0000 UTC m=+3.602176084,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.012657 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fefaf876714 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.812377364 +0000 UTC m=+3.613475418,LastTimestamp:2026-02-25 13:20:45.812377364 +0000 UTC m=+3.613475418,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.016032 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fefaf9af876 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.813659766 +0000 UTC m=+3.614757850,LastTimestamp:2026-02-25 13:20:45.813659766 +0000 UTC m=+3.614757850,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.021165 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977fefb96d2ab2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.97843013 +0000 UTC m=+3.779528184,LastTimestamp:2026-02-25 13:20:45.97843013 +0000 UTC m=+3.779528184,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.024962 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fefbae7733d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:46.003221309 +0000 UTC m=+3.804319363,LastTimestamp:2026-02-25 13:20:46.003221309 +0000 UTC m=+3.804319363,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.028455 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fefbbd24b58 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:46.018612056 +0000 UTC m=+3.819710110,LastTimestamp:2026-02-25 13:20:46.018612056 +0000 UTC m=+3.819710110,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.032120 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977fefc3e40fbc openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:46.153994172 +0000 UTC m=+3.955092226,LastTimestamp:2026-02-25 13:20:46.153994172 +0000 UTC m=+3.955092226,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.036003 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977fefc4afd189 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:46.167347593 +0000 UTC m=+3.968445647,LastTimestamp:2026-02-25 13:20:46.167347593 +0000 UTC m=+3.968445647,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.040048 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977feff590bcb7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:46.987394231 +0000 UTC m=+4.788492325,LastTimestamp:2026-02-25 13:20:46.987394231 +0000 UTC m=+4.788492325,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.044566 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977ff002b1bff2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:47.207661554 +0000 UTC m=+5.008759628,LastTimestamp:2026-02-25 13:20:47.207661554 +0000 UTC m=+5.008759628,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.048237 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977ff003448cf7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:47.217282295 +0000 UTC m=+5.018380349,LastTimestamp:2026-02-25 13:20:47.217282295 +0000 UTC m=+5.018380349,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.052156 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977ff00353c84e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:47.218280526 +0000 UTC m=+5.019378580,LastTimestamp:2026-02-25 13:20:47.218280526 +0000 UTC m=+5.019378580,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.055871 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977ff0112e46c0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:47.450703552 +0000 UTC m=+5.251801606,LastTimestamp:2026-02-25 13:20:47.450703552 +0000 UTC m=+5.251801606,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.059822 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977ff011ea9c75 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:47.463046261 +0000 UTC m=+5.264144355,LastTimestamp:2026-02-25 13:20:47.463046261 +0000 UTC m=+5.264144355,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.064236 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977ff011fd0471 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:47.464252529 +0000 UTC m=+5.265350623,LastTimestamp:2026-02-25 13:20:47.464252529 +0000 UTC m=+5.265350623,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.068362 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977ff01c78466f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:47.640102511 +0000 UTC m=+5.441200565,LastTimestamp:2026-02-25 13:20:47.640102511 +0000 UTC m=+5.441200565,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.072572 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977ff01d16e1ee openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:47.650497006 +0000 UTC m=+5.451595050,LastTimestamp:2026-02-25 13:20:47.650497006 +0000 UTC m=+5.451595050,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.076576 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977ff01d260cad openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:47.651490989 +0000 UTC m=+5.452589043,LastTimestamp:2026-02-25 13:20:47.651490989 +0000 UTC m=+5.452589043,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.080823 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977ff028eda6f2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:47.849121522 +0000 UTC m=+5.650219606,LastTimestamp:2026-02-25 13:20:47.849121522 +0000 UTC m=+5.650219606,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.085801 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977ff029b95ac3 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:47.862471363 +0000 UTC m=+5.663569467,LastTimestamp:2026-02-25 13:20:47.862471363 +0000 UTC m=+5.663569467,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.090311 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977ff029ccd112 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:47.863746834 +0000 UTC m=+5.664844888,LastTimestamp:2026-02-25 13:20:47.863746834 +0000 UTC m=+5.664844888,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.094093 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977ff034de2ef0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:48.049434352 +0000 UTC m=+5.850532406,LastTimestamp:2026-02-25 13:20:48.049434352 +0000 UTC m=+5.850532406,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.097403 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18977ff035cd27a4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:48.065095588 +0000 UTC m=+5.866193642,LastTimestamp:2026-02-25 13:20:48.065095588 +0000 UTC m=+5.866193642,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.102716 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 25 13:21:27 crc kubenswrapper[4815]: &Event{ObjectMeta:{kube-controller-manager-crc.18977ff0dcca9c54 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Feb 25 13:21:27 crc kubenswrapper[4815]: body: Feb 25 13:21:27 crc kubenswrapper[4815]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:50.866723924 +0000 UTC m=+8.667821978,LastTimestamp:2026-02-25 13:20:50.866723924 +0000 UTC m=+8.667821978,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 25 13:21:27 crc kubenswrapper[4815]: > Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.107574 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977ff0dccc4417 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:50.866832407 +0000 UTC m=+8.667930461,LastTimestamp:2026-02-25 13:20:50.866832407 +0000 UTC m=+8.667930461,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.112333 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 25 13:21:27 crc kubenswrapper[4815]: &Event{ObjectMeta:{kube-apiserver-crc.18977ff235872493 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Feb 25 13:21:27 crc kubenswrapper[4815]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Feb 25 13:21:27 crc kubenswrapper[4815]: Feb 25 13:21:27 crc kubenswrapper[4815]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:56.650441875 +0000 UTC m=+14.451539959,LastTimestamp:2026-02-25 13:20:56.650441875 +0000 UTC m=+14.451539959,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 25 13:21:27 crc kubenswrapper[4815]: > Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.115800 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977ff23588e3b3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:56.650556339 +0000 UTC m=+14.451654433,LastTimestamp:2026-02-25 13:20:56.650556339 +0000 UTC m=+14.451654433,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.121303 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.18977ff235872493\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 25 13:21:27 crc kubenswrapper[4815]: &Event{ObjectMeta:{kube-apiserver-crc.18977ff235872493 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Feb 25 13:21:27 crc kubenswrapper[4815]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Feb 25 13:21:27 crc kubenswrapper[4815]: Feb 25 13:21:27 crc kubenswrapper[4815]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:56.650441875 +0000 UTC m=+14.451539959,LastTimestamp:2026-02-25 13:20:56.657715051 +0000 UTC m=+14.458813145,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 25 13:21:27 crc kubenswrapper[4815]: > Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.126301 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.18977ff23588e3b3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977ff23588e3b3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:56.650556339 +0000 UTC m=+14.451654433,LastTimestamp:2026-02-25 13:20:56.657754762 +0000 UTC m=+14.458852846,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.130923 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.18977fefaf9af876\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fefaf9af876 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:45.813659766 +0000 UTC m=+3.614757850,LastTimestamp:2026-02-25 13:20:57.034996708 +0000 UTC m=+14.836094762,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.135323 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.18977fefbae7733d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fefbae7733d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:46.003221309 +0000 UTC m=+3.804319363,LastTimestamp:2026-02-25 13:20:57.246127687 +0000 UTC m=+15.047225741,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.139323 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.18977fefbbd24b58\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18977fefbbd24b58 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:46.018612056 +0000 UTC m=+3.819710110,LastTimestamp:2026-02-25 13:20:57.254777416 +0000 UTC m=+15.055875470,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.143234 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 25 13:21:27 crc kubenswrapper[4815]: &Event{ObjectMeta:{kube-controller-manager-crc.18977ff330e1418d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 25 13:21:27 crc kubenswrapper[4815]: body: Feb 25 13:21:27 crc kubenswrapper[4815]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:21:00.867428749 +0000 UTC m=+18.668526843,LastTimestamp:2026-02-25 13:21:00.867428749 +0000 UTC m=+18.668526843,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 25 13:21:27 crc kubenswrapper[4815]: > Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.146691 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977ff330e2291f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:21:00.867488031 +0000 UTC m=+18.668586115,LastTimestamp:2026-02-25 13:21:00.867488031 +0000 UTC m=+18.668586115,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.151831 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18977ff330e1418d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 25 13:21:27 crc kubenswrapper[4815]: &Event{ObjectMeta:{kube-controller-manager-crc.18977ff330e1418d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 25 13:21:27 crc kubenswrapper[4815]: body: Feb 25 13:21:27 crc kubenswrapper[4815]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:21:00.867428749 +0000 UTC m=+18.668526843,LastTimestamp:2026-02-25 13:21:10.867732339 +0000 UTC m=+28.668830393,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 25 13:21:27 crc kubenswrapper[4815]: > Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.155275 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18977ff330e2291f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977ff330e2291f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:21:00.867488031 +0000 UTC m=+18.668586115,LastTimestamp:2026-02-25 13:21:10.86779442 +0000 UTC m=+28.668892474,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.158847 4815 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977ff58514edef openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:21:10.870035951 +0000 UTC m=+28.671134005,LastTimestamp:2026-02-25 13:21:10.870035951 +0000 UTC m=+28.671134005,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.161945 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18977fef462cdfac\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977fef462cdfac openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.04483678 +0000 UTC m=+1.845934834,LastTimestamp:2026-02-25 13:21:10.999353325 +0000 UTC m=+28.800451379,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.165435 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18977fef565be9aa\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977fef565be9aa openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.316354986 +0000 UTC m=+2.117453040,LastTimestamp:2026-02-25 13:21:11.170433322 +0000 UTC m=+28.971531406,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.168710 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18977fef56f01c82\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977fef56f01c82 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:20:44.32606733 +0000 UTC m=+2.127165384,LastTimestamp:2026-02-25 13:21:11.181114777 +0000 UTC m=+28.982212841,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.173043 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18977ff330e1418d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 25 13:21:27 crc kubenswrapper[4815]: &Event{ObjectMeta:{kube-controller-manager-crc.18977ff330e1418d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 25 13:21:27 crc kubenswrapper[4815]: body: Feb 25 13:21:27 crc kubenswrapper[4815]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:21:00.867428749 +0000 UTC m=+18.668526843,LastTimestamp:2026-02-25 13:21:20.868367054 +0000 UTC m=+38.669465148,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 25 13:21:27 crc kubenswrapper[4815]: > Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.176133 4815 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18977ff330e2291f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18977ff330e2291f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:21:00.867488031 +0000 UTC m=+18.668586115,LastTimestamp:2026-02-25 13:21:20.868423485 +0000 UTC m=+38.669521559,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:21:27 crc kubenswrapper[4815]: W0225 13:21:27.582675 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Feb 25 13:21:27 crc kubenswrapper[4815]: E0225 13:21:27.582741 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Feb 25 13:21:27 crc kubenswrapper[4815]: I0225 13:21:27.867336 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:28 crc kubenswrapper[4815]: I0225 13:21:28.868850 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:29 crc kubenswrapper[4815]: W0225 13:21:29.305236 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Feb 25 13:21:29 crc kubenswrapper[4815]: E0225 13:21:29.305288 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Feb 25 13:21:29 crc kubenswrapper[4815]: W0225 13:21:29.507836 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:29 crc kubenswrapper[4815]: E0225 13:21:29.507914 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Feb 25 13:21:29 crc kubenswrapper[4815]: I0225 13:21:29.867138 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:30 crc kubenswrapper[4815]: I0225 13:21:30.788699 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:21:30 crc kubenswrapper[4815]: I0225 13:21:30.788860 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:30 crc kubenswrapper[4815]: I0225 13:21:30.789901 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:30 crc kubenswrapper[4815]: I0225 13:21:30.789933 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:30 crc kubenswrapper[4815]: I0225 13:21:30.789944 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:30 crc kubenswrapper[4815]: I0225 13:21:30.793595 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:21:30 crc kubenswrapper[4815]: I0225 13:21:30.865156 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:30 crc kubenswrapper[4815]: I0225 13:21:30.935368 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:30 crc kubenswrapper[4815]: I0225 13:21:30.937065 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:30 crc kubenswrapper[4815]: I0225 13:21:30.937119 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:30 crc kubenswrapper[4815]: I0225 13:21:30.937138 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:30 crc kubenswrapper[4815]: I0225 13:21:30.937953 4815 scope.go:117] "RemoveContainer" containerID="6656ddebba1eb92074357b7bfede1d0801bdb8773276da7f5e24f6d7d3a34408" Feb 25 13:21:31 crc kubenswrapper[4815]: I0225 13:21:31.081899 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:31 crc kubenswrapper[4815]: E0225 13:21:31.082279 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 25 13:21:31 crc kubenswrapper[4815]: I0225 13:21:31.083576 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:31 crc kubenswrapper[4815]: I0225 13:21:31.083640 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:31 crc kubenswrapper[4815]: I0225 13:21:31.083656 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:31 crc kubenswrapper[4815]: I0225 13:21:31.083687 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 13:21:31 crc kubenswrapper[4815]: E0225 13:21:31.088870 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 25 13:21:31 crc kubenswrapper[4815]: I0225 13:21:31.176745 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 25 13:21:31 crc kubenswrapper[4815]: I0225 13:21:31.180630 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:31 crc kubenswrapper[4815]: I0225 13:21:31.180824 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0bb59977532b467793fb68d57745c41d814b3da95fcaef8a163ec0c6cc34ce02"} Feb 25 13:21:31 crc kubenswrapper[4815]: I0225 13:21:31.181162 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:31 crc kubenswrapper[4815]: I0225 13:21:31.182391 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:31 crc kubenswrapper[4815]: I0225 13:21:31.182434 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:31 crc kubenswrapper[4815]: I0225 13:21:31.182449 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:31 crc kubenswrapper[4815]: I0225 13:21:31.182603 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:31 crc kubenswrapper[4815]: I0225 13:21:31.182641 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:31 crc kubenswrapper[4815]: I0225 13:21:31.182660 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:31 crc kubenswrapper[4815]: I0225 13:21:31.864867 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:32 crc kubenswrapper[4815]: I0225 13:21:32.401478 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 13:21:32 crc kubenswrapper[4815]: I0225 13:21:32.401773 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:32 crc kubenswrapper[4815]: I0225 13:21:32.403353 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:32 crc kubenswrapper[4815]: I0225 13:21:32.403418 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:32 crc kubenswrapper[4815]: I0225 13:21:32.403445 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:32 crc kubenswrapper[4815]: I0225 13:21:32.868136 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:33 crc kubenswrapper[4815]: E0225 13:21:33.029895 4815 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 13:21:33 crc kubenswrapper[4815]: I0225 13:21:33.189095 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 25 13:21:33 crc kubenswrapper[4815]: I0225 13:21:33.190143 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 25 13:21:33 crc kubenswrapper[4815]: I0225 13:21:33.192606 4815 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0bb59977532b467793fb68d57745c41d814b3da95fcaef8a163ec0c6cc34ce02" exitCode=255 Feb 25 13:21:33 crc kubenswrapper[4815]: I0225 13:21:33.192653 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"0bb59977532b467793fb68d57745c41d814b3da95fcaef8a163ec0c6cc34ce02"} Feb 25 13:21:33 crc kubenswrapper[4815]: I0225 13:21:33.192706 4815 scope.go:117] "RemoveContainer" containerID="6656ddebba1eb92074357b7bfede1d0801bdb8773276da7f5e24f6d7d3a34408" Feb 25 13:21:33 crc kubenswrapper[4815]: I0225 13:21:33.192941 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:33 crc kubenswrapper[4815]: I0225 13:21:33.194337 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:33 crc kubenswrapper[4815]: I0225 13:21:33.194373 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:33 crc kubenswrapper[4815]: I0225 13:21:33.194384 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:33 crc kubenswrapper[4815]: I0225 13:21:33.196224 4815 scope.go:117] "RemoveContainer" containerID="0bb59977532b467793fb68d57745c41d814b3da95fcaef8a163ec0c6cc34ce02" Feb 25 13:21:33 crc kubenswrapper[4815]: E0225 13:21:33.196642 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 13:21:33 crc kubenswrapper[4815]: I0225 13:21:33.866708 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:34 crc kubenswrapper[4815]: I0225 13:21:34.197732 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 25 13:21:34 crc kubenswrapper[4815]: I0225 13:21:34.865329 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:35 crc kubenswrapper[4815]: I0225 13:21:35.863590 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:36 crc kubenswrapper[4815]: I0225 13:21:36.864543 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:37 crc kubenswrapper[4815]: I0225 13:21:37.864189 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:38 crc kubenswrapper[4815]: E0225 13:21:38.088020 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 25 13:21:38 crc kubenswrapper[4815]: I0225 13:21:38.089100 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:38 crc kubenswrapper[4815]: I0225 13:21:38.090638 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:38 crc kubenswrapper[4815]: I0225 13:21:38.090674 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:38 crc kubenswrapper[4815]: I0225 13:21:38.090683 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:38 crc kubenswrapper[4815]: I0225 13:21:38.090701 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 13:21:38 crc kubenswrapper[4815]: E0225 13:21:38.094177 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 25 13:21:38 crc kubenswrapper[4815]: I0225 13:21:38.882038 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:21:38 crc kubenswrapper[4815]: I0225 13:21:38.882245 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:38 crc kubenswrapper[4815]: I0225 13:21:38.886975 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:38 crc kubenswrapper[4815]: I0225 13:21:38.887039 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:38 crc kubenswrapper[4815]: I0225 13:21:38.887051 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:38 crc kubenswrapper[4815]: I0225 13:21:38.887671 4815 scope.go:117] "RemoveContainer" containerID="0bb59977532b467793fb68d57745c41d814b3da95fcaef8a163ec0c6cc34ce02" Feb 25 13:21:38 crc kubenswrapper[4815]: E0225 13:21:38.887860 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 13:21:38 crc kubenswrapper[4815]: I0225 13:21:38.889552 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:39 crc kubenswrapper[4815]: I0225 13:21:39.626895 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:21:39 crc kubenswrapper[4815]: I0225 13:21:39.627071 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:39 crc kubenswrapper[4815]: I0225 13:21:39.628189 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:39 crc kubenswrapper[4815]: I0225 13:21:39.628221 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:39 crc kubenswrapper[4815]: I0225 13:21:39.628233 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:39 crc kubenswrapper[4815]: I0225 13:21:39.628841 4815 scope.go:117] "RemoveContainer" containerID="0bb59977532b467793fb68d57745c41d814b3da95fcaef8a163ec0c6cc34ce02" Feb 25 13:21:39 crc kubenswrapper[4815]: E0225 13:21:39.629036 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 13:21:39 crc kubenswrapper[4815]: I0225 13:21:39.864501 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:40 crc kubenswrapper[4815]: I0225 13:21:40.865361 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:41 crc kubenswrapper[4815]: I0225 13:21:41.866424 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:42 crc kubenswrapper[4815]: I0225 13:21:42.869662 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:43 crc kubenswrapper[4815]: E0225 13:21:43.030132 4815 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 13:21:43 crc kubenswrapper[4815]: I0225 13:21:43.866058 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:44 crc kubenswrapper[4815]: I0225 13:21:44.867597 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:45 crc kubenswrapper[4815]: I0225 13:21:45.094624 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:45 crc kubenswrapper[4815]: E0225 13:21:45.096460 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 25 13:21:45 crc kubenswrapper[4815]: I0225 13:21:45.096875 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:45 crc kubenswrapper[4815]: I0225 13:21:45.096949 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:45 crc kubenswrapper[4815]: I0225 13:21:45.096969 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:45 crc kubenswrapper[4815]: I0225 13:21:45.097050 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 13:21:45 crc kubenswrapper[4815]: E0225 13:21:45.104587 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 25 13:21:45 crc kubenswrapper[4815]: I0225 13:21:45.868637 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:46 crc kubenswrapper[4815]: I0225 13:21:46.865705 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:47 crc kubenswrapper[4815]: I0225 13:21:47.857305 4815 csr.go:261] certificate signing request csr-2kr9z is approved, waiting to be issued Feb 25 13:21:47 crc kubenswrapper[4815]: I0225 13:21:47.866573 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 13:21:47 crc kubenswrapper[4815]: I0225 13:21:47.869222 4815 csr.go:257] certificate signing request csr-2kr9z is issued Feb 25 13:21:47 crc kubenswrapper[4815]: I0225 13:21:47.884734 4815 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 25 13:21:48 crc kubenswrapper[4815]: I0225 13:21:48.710086 4815 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 25 13:21:48 crc kubenswrapper[4815]: I0225 13:21:48.870590 4815 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-03 13:32:34.036102948 +0000 UTC Feb 25 13:21:48 crc kubenswrapper[4815]: I0225 13:21:48.870631 4815 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6744h10m45.165474678s for next certificate rotation Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.105356 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.107013 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.107052 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.107062 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.107142 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.117731 4815 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.118130 4815 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 25 13:21:52 crc kubenswrapper[4815]: E0225 13:21:52.118172 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.122617 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.122649 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.122659 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.122675 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.122685 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:21:52Z","lastTransitionTime":"2026-02-25T13:21:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:21:52 crc kubenswrapper[4815]: E0225 13:21:52.135742 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c5638ab-7489-4aa7-9d40-f8fadb68ce90\\\",\\\"systemUUID\\\":\\\"86509d21-e518-46f8-a810-5a92ff458924\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.143652 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.143700 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.143717 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.143742 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.143761 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:21:52Z","lastTransitionTime":"2026-02-25T13:21:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:21:52 crc kubenswrapper[4815]: E0225 13:21:52.153263 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c5638ab-7489-4aa7-9d40-f8fadb68ce90\\\",\\\"systemUUID\\\":\\\"86509d21-e518-46f8-a810-5a92ff458924\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.161442 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.161485 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.161497 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.161563 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.161587 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:21:52Z","lastTransitionTime":"2026-02-25T13:21:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:21:52 crc kubenswrapper[4815]: E0225 13:21:52.174038 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c5638ab-7489-4aa7-9d40-f8fadb68ce90\\\",\\\"systemUUID\\\":\\\"86509d21-e518-46f8-a810-5a92ff458924\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.180914 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.180957 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.180972 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.180990 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:21:52 crc kubenswrapper[4815]: I0225 13:21:52.181047 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:21:52Z","lastTransitionTime":"2026-02-25T13:21:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:21:52 crc kubenswrapper[4815]: E0225 13:21:52.192704 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:21:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c5638ab-7489-4aa7-9d40-f8fadb68ce90\\\",\\\"systemUUID\\\":\\\"86509d21-e518-46f8-a810-5a92ff458924\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 13:21:52 crc kubenswrapper[4815]: E0225 13:21:52.192865 4815 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 25 13:21:52 crc kubenswrapper[4815]: E0225 13:21:52.192900 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:52 crc kubenswrapper[4815]: E0225 13:21:52.293544 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:52 crc kubenswrapper[4815]: E0225 13:21:52.394209 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:52 crc kubenswrapper[4815]: E0225 13:21:52.494623 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:52 crc kubenswrapper[4815]: E0225 13:21:52.595296 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:52 crc kubenswrapper[4815]: E0225 13:21:52.696613 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:52 crc kubenswrapper[4815]: E0225 13:21:52.797137 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:52 crc kubenswrapper[4815]: E0225 13:21:52.897685 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:52 crc kubenswrapper[4815]: E0225 13:21:52.997983 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:53 crc kubenswrapper[4815]: E0225 13:21:53.030277 4815 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 13:21:53 crc kubenswrapper[4815]: E0225 13:21:53.098389 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:53 crc kubenswrapper[4815]: E0225 13:21:53.199043 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:53 crc kubenswrapper[4815]: E0225 13:21:53.299139 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:53 crc kubenswrapper[4815]: E0225 13:21:53.399738 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:53 crc kubenswrapper[4815]: E0225 13:21:53.500425 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:53 crc kubenswrapper[4815]: E0225 13:21:53.601527 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:53 crc kubenswrapper[4815]: E0225 13:21:53.702498 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:53 crc kubenswrapper[4815]: E0225 13:21:53.802891 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:53 crc kubenswrapper[4815]: E0225 13:21:53.903268 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:54 crc kubenswrapper[4815]: E0225 13:21:54.003391 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:54 crc kubenswrapper[4815]: I0225 13:21:54.059912 4815 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 25 13:21:54 crc kubenswrapper[4815]: E0225 13:21:54.103554 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:54 crc kubenswrapper[4815]: E0225 13:21:54.204337 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:54 crc kubenswrapper[4815]: E0225 13:21:54.304654 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:54 crc kubenswrapper[4815]: E0225 13:21:54.405713 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:54 crc kubenswrapper[4815]: E0225 13:21:54.506376 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:54 crc kubenswrapper[4815]: E0225 13:21:54.607416 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:54 crc kubenswrapper[4815]: E0225 13:21:54.708694 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:54 crc kubenswrapper[4815]: E0225 13:21:54.809738 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:54 crc kubenswrapper[4815]: E0225 13:21:54.910209 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:54 crc kubenswrapper[4815]: I0225 13:21:54.934916 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 13:21:54 crc kubenswrapper[4815]: I0225 13:21:54.936540 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:21:54 crc kubenswrapper[4815]: I0225 13:21:54.936564 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:21:54 crc kubenswrapper[4815]: I0225 13:21:54.936582 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:21:54 crc kubenswrapper[4815]: I0225 13:21:54.937095 4815 scope.go:117] "RemoveContainer" containerID="0bb59977532b467793fb68d57745c41d814b3da95fcaef8a163ec0c6cc34ce02" Feb 25 13:21:54 crc kubenswrapper[4815]: E0225 13:21:54.937274 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 13:21:55 crc kubenswrapper[4815]: E0225 13:21:55.010458 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:55 crc kubenswrapper[4815]: E0225 13:21:55.111528 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:55 crc kubenswrapper[4815]: E0225 13:21:55.212467 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:55 crc kubenswrapper[4815]: E0225 13:21:55.313563 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:55 crc kubenswrapper[4815]: E0225 13:21:55.413996 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:55 crc kubenswrapper[4815]: E0225 13:21:55.514450 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:55 crc kubenswrapper[4815]: E0225 13:21:55.614733 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:55 crc kubenswrapper[4815]: E0225 13:21:55.715622 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:55 crc kubenswrapper[4815]: E0225 13:21:55.816780 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:55 crc kubenswrapper[4815]: E0225 13:21:55.917682 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:56 crc kubenswrapper[4815]: E0225 13:21:56.018142 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:56 crc kubenswrapper[4815]: E0225 13:21:56.118755 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:56 crc kubenswrapper[4815]: E0225 13:21:56.219280 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:56 crc kubenswrapper[4815]: E0225 13:21:56.319755 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:56 crc kubenswrapper[4815]: E0225 13:21:56.420455 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:56 crc kubenswrapper[4815]: E0225 13:21:56.521032 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:56 crc kubenswrapper[4815]: E0225 13:21:56.622106 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:56 crc kubenswrapper[4815]: E0225 13:21:56.722459 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:56 crc kubenswrapper[4815]: E0225 13:21:56.823111 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:56 crc kubenswrapper[4815]: E0225 13:21:56.924306 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:57 crc kubenswrapper[4815]: E0225 13:21:57.025402 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:57 crc kubenswrapper[4815]: E0225 13:21:57.125975 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:57 crc kubenswrapper[4815]: E0225 13:21:57.226400 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:57 crc kubenswrapper[4815]: E0225 13:21:57.326778 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:57 crc kubenswrapper[4815]: E0225 13:21:57.427337 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:57 crc kubenswrapper[4815]: E0225 13:21:57.527763 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:57 crc kubenswrapper[4815]: E0225 13:21:57.628823 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:57 crc kubenswrapper[4815]: E0225 13:21:57.729504 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:57 crc kubenswrapper[4815]: E0225 13:21:57.830155 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:57 crc kubenswrapper[4815]: E0225 13:21:57.931125 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:58 crc kubenswrapper[4815]: E0225 13:21:58.032260 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:58 crc kubenswrapper[4815]: E0225 13:21:58.132539 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:58 crc kubenswrapper[4815]: E0225 13:21:58.232694 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:58 crc kubenswrapper[4815]: E0225 13:21:58.333238 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:58 crc kubenswrapper[4815]: E0225 13:21:58.433854 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:58 crc kubenswrapper[4815]: E0225 13:21:58.534033 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:58 crc kubenswrapper[4815]: E0225 13:21:58.635169 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:58 crc kubenswrapper[4815]: E0225 13:21:58.735818 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:58 crc kubenswrapper[4815]: E0225 13:21:58.836894 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:58 crc kubenswrapper[4815]: E0225 13:21:58.937593 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:59 crc kubenswrapper[4815]: E0225 13:21:59.038555 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:59 crc kubenswrapper[4815]: E0225 13:21:59.139711 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:59 crc kubenswrapper[4815]: E0225 13:21:59.239882 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:59 crc kubenswrapper[4815]: E0225 13:21:59.340337 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:59 crc kubenswrapper[4815]: E0225 13:21:59.440600 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:59 crc kubenswrapper[4815]: E0225 13:21:59.541986 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:59 crc kubenswrapper[4815]: E0225 13:21:59.643268 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:59 crc kubenswrapper[4815]: E0225 13:21:59.744217 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:59 crc kubenswrapper[4815]: E0225 13:21:59.845152 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:21:59 crc kubenswrapper[4815]: E0225 13:21:59.946351 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:00 crc kubenswrapper[4815]: E0225 13:22:00.046950 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:00 crc kubenswrapper[4815]: E0225 13:22:00.148210 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:00 crc kubenswrapper[4815]: E0225 13:22:00.248685 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:00 crc kubenswrapper[4815]: E0225 13:22:00.350412 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:00 crc kubenswrapper[4815]: E0225 13:22:00.450944 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:00 crc kubenswrapper[4815]: E0225 13:22:00.551409 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:00 crc kubenswrapper[4815]: E0225 13:22:00.651928 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:00 crc kubenswrapper[4815]: E0225 13:22:00.752638 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:00 crc kubenswrapper[4815]: E0225 13:22:00.852921 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:00 crc kubenswrapper[4815]: E0225 13:22:00.953395 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:01 crc kubenswrapper[4815]: E0225 13:22:01.054801 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:01 crc kubenswrapper[4815]: E0225 13:22:01.155673 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:01 crc kubenswrapper[4815]: E0225 13:22:01.255778 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:01 crc kubenswrapper[4815]: E0225 13:22:01.356192 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:01 crc kubenswrapper[4815]: E0225 13:22:01.457273 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:01 crc kubenswrapper[4815]: E0225 13:22:01.558409 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:01 crc kubenswrapper[4815]: E0225 13:22:01.659284 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:01 crc kubenswrapper[4815]: E0225 13:22:01.759653 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:01 crc kubenswrapper[4815]: E0225 13:22:01.860473 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:01 crc kubenswrapper[4815]: E0225 13:22:01.961697 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:02 crc kubenswrapper[4815]: E0225 13:22:02.062544 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:02 crc kubenswrapper[4815]: E0225 13:22:02.162685 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:02 crc kubenswrapper[4815]: E0225 13:22:02.263169 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:02 crc kubenswrapper[4815]: E0225 13:22:02.363531 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:02 crc kubenswrapper[4815]: E0225 13:22:02.464308 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:02 crc kubenswrapper[4815]: E0225 13:22:02.491908 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.496914 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.496939 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.496952 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.496967 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.496978 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:02Z","lastTransitionTime":"2026-02-25T13:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:02 crc kubenswrapper[4815]: E0225 13:22:02.511928 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c5638ab-7489-4aa7-9d40-f8fadb68ce90\\\",\\\"systemUUID\\\":\\\"86509d21-e518-46f8-a810-5a92ff458924\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.516303 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.516363 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.516385 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.516411 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.516432 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:02Z","lastTransitionTime":"2026-02-25T13:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:02 crc kubenswrapper[4815]: E0225 13:22:02.527229 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c5638ab-7489-4aa7-9d40-f8fadb68ce90\\\",\\\"systemUUID\\\":\\\"86509d21-e518-46f8-a810-5a92ff458924\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.531425 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.531472 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.531490 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.531538 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.531558 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:02Z","lastTransitionTime":"2026-02-25T13:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:02 crc kubenswrapper[4815]: E0225 13:22:02.544845 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c5638ab-7489-4aa7-9d40-f8fadb68ce90\\\",\\\"systemUUID\\\":\\\"86509d21-e518-46f8-a810-5a92ff458924\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.548177 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.548225 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.548245 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.548270 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:02 crc kubenswrapper[4815]: I0225 13:22:02.548288 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:02Z","lastTransitionTime":"2026-02-25T13:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:02 crc kubenswrapper[4815]: E0225 13:22:02.564784 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0c5638ab-7489-4aa7-9d40-f8fadb68ce90\\\",\\\"systemUUID\\\":\\\"86509d21-e518-46f8-a810-5a92ff458924\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 13:22:02 crc kubenswrapper[4815]: E0225 13:22:02.565119 4815 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 25 13:22:02 crc kubenswrapper[4815]: E0225 13:22:02.565244 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:02 crc kubenswrapper[4815]: E0225 13:22:02.665864 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:02 crc kubenswrapper[4815]: E0225 13:22:02.766002 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:02 crc kubenswrapper[4815]: E0225 13:22:02.867289 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:02 crc kubenswrapper[4815]: E0225 13:22:02.968071 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:03 crc kubenswrapper[4815]: E0225 13:22:03.030434 4815 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 13:22:03 crc kubenswrapper[4815]: E0225 13:22:03.068484 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:03 crc kubenswrapper[4815]: E0225 13:22:03.169479 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:03 crc kubenswrapper[4815]: E0225 13:22:03.269647 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:03 crc kubenswrapper[4815]: E0225 13:22:03.370612 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:03 crc kubenswrapper[4815]: E0225 13:22:03.470762 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:03 crc kubenswrapper[4815]: E0225 13:22:03.571944 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:03 crc kubenswrapper[4815]: E0225 13:22:03.673041 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:03 crc kubenswrapper[4815]: E0225 13:22:03.773863 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:03 crc kubenswrapper[4815]: E0225 13:22:03.874360 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:03 crc kubenswrapper[4815]: E0225 13:22:03.974870 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:04 crc kubenswrapper[4815]: E0225 13:22:04.075424 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:04 crc kubenswrapper[4815]: E0225 13:22:04.176156 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:04 crc kubenswrapper[4815]: E0225 13:22:04.276772 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:04 crc kubenswrapper[4815]: E0225 13:22:04.377692 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:04 crc kubenswrapper[4815]: E0225 13:22:04.478018 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:04 crc kubenswrapper[4815]: E0225 13:22:04.578331 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:04 crc kubenswrapper[4815]: E0225 13:22:04.678827 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:04 crc kubenswrapper[4815]: E0225 13:22:04.779132 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:04 crc kubenswrapper[4815]: E0225 13:22:04.880158 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:04 crc kubenswrapper[4815]: E0225 13:22:04.981328 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:05 crc kubenswrapper[4815]: E0225 13:22:05.081489 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:05 crc kubenswrapper[4815]: E0225 13:22:05.182566 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:05 crc kubenswrapper[4815]: E0225 13:22:05.282704 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:05 crc kubenswrapper[4815]: E0225 13:22:05.383688 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:05 crc kubenswrapper[4815]: E0225 13:22:05.484110 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:05 crc kubenswrapper[4815]: I0225 13:22:05.507580 4815 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 25 13:22:05 crc kubenswrapper[4815]: E0225 13:22:05.584896 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:05 crc kubenswrapper[4815]: E0225 13:22:05.685383 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:05 crc kubenswrapper[4815]: E0225 13:22:05.785794 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:05 crc kubenswrapper[4815]: E0225 13:22:05.886631 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:05 crc kubenswrapper[4815]: E0225 13:22:05.987194 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:06 crc kubenswrapper[4815]: E0225 13:22:06.087934 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:06 crc kubenswrapper[4815]: E0225 13:22:06.189080 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:06 crc kubenswrapper[4815]: E0225 13:22:06.289924 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:06 crc kubenswrapper[4815]: E0225 13:22:06.390088 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:06 crc kubenswrapper[4815]: E0225 13:22:06.491078 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:06 crc kubenswrapper[4815]: E0225 13:22:06.592238 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:06 crc kubenswrapper[4815]: E0225 13:22:06.693189 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:06 crc kubenswrapper[4815]: E0225 13:22:06.794079 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:06 crc kubenswrapper[4815]: E0225 13:22:06.895012 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 13:22:06 crc kubenswrapper[4815]: I0225 13:22:06.956439 4815 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 25 13:22:06 crc kubenswrapper[4815]: I0225 13:22:06.998397 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:06 crc kubenswrapper[4815]: I0225 13:22:06.998543 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:06 crc kubenswrapper[4815]: I0225 13:22:06.998572 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:06 crc kubenswrapper[4815]: I0225 13:22:06.998603 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:06 crc kubenswrapper[4815]: I0225 13:22:06.998626 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:06Z","lastTransitionTime":"2026-02-25T13:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.102220 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.102595 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.102811 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.103028 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.103208 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:07Z","lastTransitionTime":"2026-02-25T13:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.207565 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.207642 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.207661 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.207688 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.207707 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:07Z","lastTransitionTime":"2026-02-25T13:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.310178 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.310262 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.310289 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.310321 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.310343 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:07Z","lastTransitionTime":"2026-02-25T13:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.414159 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.414228 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.414250 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.414281 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.414303 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:07Z","lastTransitionTime":"2026-02-25T13:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.517154 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.517214 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.517232 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.517258 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.517278 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:07Z","lastTransitionTime":"2026-02-25T13:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.620462 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.620499 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.620528 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.620544 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.620557 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:07Z","lastTransitionTime":"2026-02-25T13:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.723701 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.723758 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.723774 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.723794 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.723808 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:07Z","lastTransitionTime":"2026-02-25T13:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.826721 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.826999 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.827075 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.827174 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.827263 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:07Z","lastTransitionTime":"2026-02-25T13:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.898014 4815 apiserver.go:52] "Watching apiserver" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.902720 4815 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.903097 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.903774 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.903819 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:07 crc kubenswrapper[4815]: E0225 13:22:07.903897 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.903958 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.904205 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:07 crc kubenswrapper[4815]: E0225 13:22:07.904298 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.904374 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 13:22:07 crc kubenswrapper[4815]: E0225 13:22:07.904407 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.903805 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.906312 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.907353 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.907593 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.908213 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.908628 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.908681 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.908714 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.908722 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.908765 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.929786 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.929835 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.929848 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.929872 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.929885 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:07Z","lastTransitionTime":"2026-02-25T13:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.936519 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.948690 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.960140 4815 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.964301 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.978906 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 13:22:07 crc kubenswrapper[4815]: I0225 13:22:07.996427 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003303 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003369 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003404 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003474 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003536 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003571 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003600 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003634 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003665 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003692 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003724 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003753 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003788 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003818 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003837 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003848 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003927 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003962 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.003994 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004027 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004056 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004086 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004116 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004147 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004176 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004209 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004235 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004243 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004310 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004343 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004378 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004373 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004409 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004460 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004547 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004698 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004759 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004811 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004864 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004914 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004986 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005041 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005087 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005130 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005175 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005224 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005285 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005339 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005393 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005442 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005496 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005592 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005643 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005693 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005745 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005793 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005852 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005901 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005949 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.006011 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.006057 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004599 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.006102 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.006152 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.006202 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.006249 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.006297 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.006345 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.006397 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.006455 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.006532 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008574 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008627 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008652 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008676 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008700 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008725 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008747 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008772 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008796 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008820 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008849 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008875 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008900 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008954 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008981 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009006 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009029 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009053 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009076 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009101 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009125 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009145 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009167 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009191 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009215 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009240 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009267 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009291 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009315 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009340 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009364 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009387 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009409 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009435 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009459 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009483 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009521 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009549 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009571 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009593 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009617 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009642 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009674 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009699 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009723 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009747 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009770 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009792 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009818 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009843 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009867 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009893 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009915 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009939 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009964 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009992 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.010016 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.010041 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.010066 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.010089 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.010112 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.010137 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.010162 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004687 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004782 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.004954 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005143 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005209 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005256 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005313 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005394 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005646 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005729 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005727 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005835 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005925 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.005934 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.006063 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.006083 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.006100 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.006262 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.006775 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.010897 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.010936 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.010961 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.010985 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011008 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011036 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011061 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011086 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011139 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011243 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011271 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011299 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011322 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011346 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011371 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011396 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011419 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011482 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011528 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011555 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011583 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011610 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011637 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011663 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011690 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011717 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011742 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011772 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011805 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011834 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011859 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011887 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011921 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011949 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011976 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012000 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012023 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012046 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012068 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012090 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012114 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012140 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012166 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012194 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012220 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012246 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012272 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012295 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012318 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012341 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012367 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012394 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012421 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012446 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012471 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012497 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012542 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012567 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012591 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012621 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012646 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012673 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012719 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012758 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012787 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012848 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012886 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012917 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012944 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012971 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013000 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013025 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013050 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013078 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013105 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013135 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013158 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013187 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013213 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013294 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013313 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013328 4815 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013342 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013355 4815 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013370 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013382 4815 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013397 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013411 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013426 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013440 4815 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013453 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013467 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013481 4815 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013494 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013524 4815 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013539 4815 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013553 4815 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013567 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013583 4815 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013597 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013611 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013997 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.014150 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.006890 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.007083 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.007145 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.007191 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.007199 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.007214 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.007333 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.007891 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008370 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008791 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.008929 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009083 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009146 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009282 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009672 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009774 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009809 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009814 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.009921 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.022720 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.010431 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.010504 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.010622 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.010703 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.010913 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.010574 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011427 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011606 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011705 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011708 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.011881 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012252 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012743 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012945 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012953 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013171 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.012487 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013369 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013410 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013466 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013518 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.013546 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.014013 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.014009 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.014064 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.014197 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.014358 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.014420 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.014632 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.015105 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.015354 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.015363 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.015400 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.015633 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.015810 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.015928 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.016150 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.016187 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.016330 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.016618 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.016607 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.016949 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.017291 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.017348 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.017940 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.018014 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.018322 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.018476 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.019047 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.019314 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.020175 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.020337 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.020369 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.020538 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.020898 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.020998 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.021026 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.021051 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.021074 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.021093 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.021111 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:08.521092667 +0000 UTC m=+86.322190721 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.021288 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.021291 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.021481 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.021621 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.021753 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.021793 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.021830 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.021963 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.022083 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.022490 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.022746 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.023144 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.023176 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.023298 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.024005 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.024011 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.024812 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.024875 4815 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.025575 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.026404 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.027175 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.027646 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.028064 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.028264 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.028306 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.028831 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.029141 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.029200 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.029403 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.029465 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.029463 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.029625 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.029652 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.029865 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.029994 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.030123 4815 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.031276 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:08.531239314 +0000 UTC m=+86.332337408 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.031281 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.031521 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.031116 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.031960 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.032001 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.030414 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.030577 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.030832 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.030939 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.032618 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.030269 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.033251 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.033538 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.034057 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.033868 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.033840 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.034372 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.034429 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.034760 4815 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.034956 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:08.534931057 +0000 UTC m=+86.336029111 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.035097 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.035651 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.034969 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.036034 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.036251 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.036422 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.036692 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.037985 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.040585 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.040714 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.040813 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.040907 4815 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.041032 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:08.541012492 +0000 UTC m=+86.342110546 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.045040 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.046573 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.047301 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.047716 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.050058 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.050097 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.050049 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.050193 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.050382 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.050625 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.050743 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.050980 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.051040 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.051064 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.051094 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.051115 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:08Z","lastTransitionTime":"2026-02-25T13:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.052880 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.053077 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.053117 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.053136 4815 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.053216 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:08.553185691 +0000 UTC m=+86.354283815 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.058593 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.059169 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.059844 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.060491 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.060969 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.061027 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.061082 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.061118 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.061137 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.061188 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.061538 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.061630 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.061746 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.061761 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.062047 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.062045 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.062058 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.062496 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.062610 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.062998 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.065629 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.068058 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.072219 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.072902 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.074035 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.087133 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.090528 4815 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.113944 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.113990 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114035 4815 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114045 4815 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114054 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114062 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114070 4815 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114079 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114088 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114095 4815 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114104 4815 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114112 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114120 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114128 4815 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114135 4815 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114143 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114150 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114158 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114166 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114174 4815 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114181 4815 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114189 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114197 4815 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114205 4815 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114213 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114220 4815 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114228 4815 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114235 4815 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114243 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114168 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114252 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114327 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114354 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114382 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114235 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114406 4815 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114569 4815 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114595 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114614 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114632 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114644 4815 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114657 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114669 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114681 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114692 4815 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114704 4815 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114714 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114726 4815 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114737 4815 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114749 4815 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114761 4815 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114772 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114784 4815 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114796 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114807 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114818 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114828 4815 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114840 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114851 4815 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114862 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114887 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114899 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114911 4815 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114922 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114933 4815 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114945 4815 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114957 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114969 4815 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114980 4815 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.114991 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115002 4815 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115014 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115024 4815 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115037 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115049 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115061 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115073 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115084 4815 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115096 4815 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115109 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115122 4815 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115133 4815 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115145 4815 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115156 4815 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115167 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115179 4815 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115189 4815 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115201 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115212 4815 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115222 4815 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115233 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115244 4815 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115256 4815 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115267 4815 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115278 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115290 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115301 4815 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115312 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115324 4815 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115335 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115346 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115357 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115369 4815 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115381 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115423 4815 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115434 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115445 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115458 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115470 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115482 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115493 4815 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115571 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115585 4815 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115595 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115607 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115618 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115629 4815 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115640 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115651 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115669 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115681 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115695 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115707 4815 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115719 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115729 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115740 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115753 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115764 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115774 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115785 4815 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115796 4815 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115807 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115817 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115829 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115840 4815 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115851 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115863 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115873 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115884 4815 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115896 4815 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115906 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115918 4815 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115929 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115940 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115951 4815 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115963 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115975 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115985 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.115996 4815 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116007 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116018 4815 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116030 4815 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116041 4815 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116052 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116063 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116075 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116086 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116098 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116109 4815 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116121 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116132 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116144 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116154 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116166 4815 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116177 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116188 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116199 4815 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116212 4815 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116223 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116234 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116245 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116256 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116267 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116278 4815 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116290 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.116302 4815 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.154188 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.154234 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.154261 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.154287 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.154304 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:08Z","lastTransitionTime":"2026-02-25T13:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.221744 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.228935 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.236740 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 13:22:08 crc kubenswrapper[4815]: W0225 13:22:08.247396 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-f5f0749d3b3fa32d50987ee6c2da83447e12529cf7acec383e78b84e32b4e420 WatchSource:0}: Error finding container f5f0749d3b3fa32d50987ee6c2da83447e12529cf7acec383e78b84e32b4e420: Status 404 returned error can't find the container with id f5f0749d3b3fa32d50987ee6c2da83447e12529cf7acec383e78b84e32b4e420 Feb 25 13:22:08 crc kubenswrapper[4815]: W0225 13:22:08.257700 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-6174f6b05a447b48cc9df0573bd2364a7f8d49dc6423e643c68d9cdc0458a291 WatchSource:0}: Error finding container 6174f6b05a447b48cc9df0573bd2364a7f8d49dc6423e643c68d9cdc0458a291: Status 404 returned error can't find the container with id 6174f6b05a447b48cc9df0573bd2364a7f8d49dc6423e643c68d9cdc0458a291 Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.258139 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.258200 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.258224 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.258253 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.258277 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:08Z","lastTransitionTime":"2026-02-25T13:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.292286 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"47e1ec1709598c8fb4393a1a0b98b1ccc2465c524731cf32a6bacb8f2d49e953"} Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.293848 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6174f6b05a447b48cc9df0573bd2364a7f8d49dc6423e643c68d9cdc0458a291"} Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.295229 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f5f0749d3b3fa32d50987ee6c2da83447e12529cf7acec383e78b84e32b4e420"} Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.361895 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.361968 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.361981 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.362000 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.362012 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:08Z","lastTransitionTime":"2026-02-25T13:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.463991 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.464028 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.464038 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.464054 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.464064 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:08Z","lastTransitionTime":"2026-02-25T13:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.524909 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.525159 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:09.525107882 +0000 UTC m=+87.326205946 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.566171 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.566210 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.566224 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.566243 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.566257 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:08Z","lastTransitionTime":"2026-02-25T13:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.626151 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.626222 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.626266 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.626314 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.626399 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.626432 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.626448 4815 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.626454 4815 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.626462 4815 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.626543 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:09.626486599 +0000 UTC m=+87.427584663 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.626577 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:09.626556481 +0000 UTC m=+87.427654575 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.626400 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.626604 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:09.626591523 +0000 UTC m=+87.427689627 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.626622 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.626642 4815 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:08 crc kubenswrapper[4815]: E0225 13:22:08.626706 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:09.626688295 +0000 UTC m=+87.427786389 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.669838 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.669911 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.669945 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.669975 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.669998 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:08Z","lastTransitionTime":"2026-02-25T13:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.772170 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.772211 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.772223 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.772264 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.772279 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:08Z","lastTransitionTime":"2026-02-25T13:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.875401 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.875472 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.875501 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.875571 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.875597 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:08Z","lastTransitionTime":"2026-02-25T13:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.942441 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.943433 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.944899 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.945693 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.946927 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.947650 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.948415 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.949716 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.951149 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.953254 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.954464 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.957396 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.958664 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.959952 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.961994 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.963212 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.965278 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.966782 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.968307 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.970753 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.971765 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.973809 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.974709 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.977470 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.978953 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.979045 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.979073 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.979100 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.979120 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:08Z","lastTransitionTime":"2026-02-25T13:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.979389 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.981129 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.983352 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.984559 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.985997 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.987458 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.988076 4815 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.988217 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.989851 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.990437 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.991128 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.992448 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.993226 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.993840 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.994551 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.995313 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.995885 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.997956 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 25 13:22:08 crc kubenswrapper[4815]: I0225 13:22:08.999647 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.001073 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.002052 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.003304 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.005817 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.007800 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.010033 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.011239 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.013191 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.014821 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.016711 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.018628 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.081783 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.081838 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.081855 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.081913 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.081931 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:09Z","lastTransitionTime":"2026-02-25T13:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.184294 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.184363 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.184375 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.184392 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.184403 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:09Z","lastTransitionTime":"2026-02-25T13:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.287875 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.287923 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.287934 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.287951 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.287962 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:09Z","lastTransitionTime":"2026-02-25T13:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.301482 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"8cac4d62ae7a1a4706481e31f5cc2f19be8f0c0ecb59c9c3cc444eb5f95e69e6"} Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.305938 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ef2854f740724f0351fcabe1b903a77478843cfb9cebba59896d2c3c4b74123a"} Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.306026 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"55b21f77970848ba60c2567e04d4989a20b0689f9b86089f7a3c3613cd8c7923"} Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.321814 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:22:09Z is after 2025-08-24T17:21:41Z" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.336291 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:22:09Z is after 2025-08-24T17:21:41Z" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.356854 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:22:09Z is after 2025-08-24T17:21:41Z" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.375768 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:22:09Z is after 2025-08-24T17:21:41Z" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.390918 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.390957 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.390969 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.391008 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.391020 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:09Z","lastTransitionTime":"2026-02-25T13:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.391426 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:22:09Z is after 2025-08-24T17:21:41Z" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.409758 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cac4d62ae7a1a4706481e31f5cc2f19be8f0c0ecb59c9c3cc444eb5f95e69e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-25T13:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:22:09Z is after 2025-08-24T17:21:41Z" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.425235 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:22:09Z is after 2025-08-24T17:21:41Z" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.438719 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:22:09Z is after 2025-08-24T17:21:41Z" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.455496 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cac4d62ae7a1a4706481e31f5cc2f19be8f0c0ecb59c9c3cc444eb5f95e69e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-25T13:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:22:09Z is after 2025-08-24T17:21:41Z" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.468382 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef2854f740724f0351fcabe1b903a77478843cfb9cebba59896d2c3c4b74123a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-25T13:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55b21f77970848ba60c2567e04d4989a20b0689f9b86089f7a3c3613cd8c7923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-25T13:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:22:09Z is after 2025-08-24T17:21:41Z" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.483657 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:22:09Z is after 2025-08-24T17:21:41Z" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.493381 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.493414 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.493423 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.493436 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.493445 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:09Z","lastTransitionTime":"2026-02-25T13:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.500369 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T13:22:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T13:22:09Z is after 2025-08-24T17:21:41Z" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.535648 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:09 crc kubenswrapper[4815]: E0225 13:22:09.536043 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:11.535821703 +0000 UTC m=+89.336919757 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.596492 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.596576 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.596590 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.596636 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.596648 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:09Z","lastTransitionTime":"2026-02-25T13:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.636379 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.636463 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:09 crc kubenswrapper[4815]: E0225 13:22:09.636500 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 13:22:09 crc kubenswrapper[4815]: E0225 13:22:09.636537 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 13:22:09 crc kubenswrapper[4815]: E0225 13:22:09.636549 4815 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:09 crc kubenswrapper[4815]: E0225 13:22:09.636598 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:11.636580093 +0000 UTC m=+89.437678177 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.636500 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:09 crc kubenswrapper[4815]: E0225 13:22:09.636639 4815 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 13:22:09 crc kubenswrapper[4815]: E0225 13:22:09.636651 4815 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.636681 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:09 crc kubenswrapper[4815]: E0225 13:22:09.636717 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:11.636673345 +0000 UTC m=+89.437771409 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 13:22:09 crc kubenswrapper[4815]: E0225 13:22:09.636745 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:11.636729747 +0000 UTC m=+89.437827821 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 13:22:09 crc kubenswrapper[4815]: E0225 13:22:09.636769 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 13:22:09 crc kubenswrapper[4815]: E0225 13:22:09.636801 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 13:22:09 crc kubenswrapper[4815]: E0225 13:22:09.636809 4815 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:09 crc kubenswrapper[4815]: E0225 13:22:09.636845 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:11.63683721 +0000 UTC m=+89.437935264 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.698456 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.698498 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.698530 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.698545 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.698558 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:09Z","lastTransitionTime":"2026-02-25T13:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.801354 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.801399 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.801408 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.801423 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.801432 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:09Z","lastTransitionTime":"2026-02-25T13:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.903789 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.903833 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.903848 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.903869 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.903884 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:09Z","lastTransitionTime":"2026-02-25T13:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.934520 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.934576 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:09 crc kubenswrapper[4815]: E0225 13:22:09.934669 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.934688 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:09 crc kubenswrapper[4815]: E0225 13:22:09.934763 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 13:22:09 crc kubenswrapper[4815]: E0225 13:22:09.934835 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.944630 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 25 13:22:09 crc kubenswrapper[4815]: I0225 13:22:09.944884 4815 scope.go:117] "RemoveContainer" containerID="0bb59977532b467793fb68d57745c41d814b3da95fcaef8a163ec0c6cc34ce02" Feb 25 13:22:09 crc kubenswrapper[4815]: E0225 13:22:09.945081 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.005944 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.005979 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.006006 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.006024 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.006035 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:10Z","lastTransitionTime":"2026-02-25T13:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.116461 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.116539 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.116551 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.116569 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.116579 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:10Z","lastTransitionTime":"2026-02-25T13:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.219593 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.219653 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.219668 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.219688 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.219705 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:10Z","lastTransitionTime":"2026-02-25T13:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.309234 4815 scope.go:117] "RemoveContainer" containerID="0bb59977532b467793fb68d57745c41d814b3da95fcaef8a163ec0c6cc34ce02" Feb 25 13:22:10 crc kubenswrapper[4815]: E0225 13:22:10.309407 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.322243 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.322272 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.322279 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.322291 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.322299 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:10Z","lastTransitionTime":"2026-02-25T13:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.424121 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.424187 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.424210 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.424238 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.424260 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:10Z","lastTransitionTime":"2026-02-25T13:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.527295 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.527348 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.527365 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.527388 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.527404 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:10Z","lastTransitionTime":"2026-02-25T13:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.629592 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.629651 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.629669 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.629695 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.629712 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:10Z","lastTransitionTime":"2026-02-25T13:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.732045 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.732079 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.732089 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.732103 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.732113 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:10Z","lastTransitionTime":"2026-02-25T13:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.833956 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.834360 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.834444 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.834534 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.834605 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:10Z","lastTransitionTime":"2026-02-25T13:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.953445 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:10 crc kubenswrapper[4815]: E0225 13:22:10.953581 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.959970 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.960001 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.960011 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.960024 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:10 crc kubenswrapper[4815]: I0225 13:22:10.960036 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:10Z","lastTransitionTime":"2026-02-25T13:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.062895 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.062956 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.062966 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.063000 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.063014 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:11Z","lastTransitionTime":"2026-02-25T13:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.166221 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.166270 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.166281 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.166298 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.166311 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:11Z","lastTransitionTime":"2026-02-25T13:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.268536 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.268604 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.268624 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.268651 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.268669 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:11Z","lastTransitionTime":"2026-02-25T13:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.314140 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8f77d7410cae9504f4a855262e8896b2d0b991b84a19dbbb25ddc2ec5c8722f2"} Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.370814 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.370884 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.370906 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.370947 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.370969 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:11Z","lastTransitionTime":"2026-02-25T13:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.473898 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.473943 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.473952 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.473968 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.473979 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:11Z","lastTransitionTime":"2026-02-25T13:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.558786 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:11 crc kubenswrapper[4815]: E0225 13:22:11.559032 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:15.558987047 +0000 UTC m=+93.360085141 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.577105 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.577179 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.577199 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.577225 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.577243 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:11Z","lastTransitionTime":"2026-02-25T13:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.659794 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.659833 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.659853 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.659873 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:11 crc kubenswrapper[4815]: E0225 13:22:11.659925 4815 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 13:22:11 crc kubenswrapper[4815]: E0225 13:22:11.659975 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 13:22:11 crc kubenswrapper[4815]: E0225 13:22:11.659996 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 13:22:11 crc kubenswrapper[4815]: E0225 13:22:11.660010 4815 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:11 crc kubenswrapper[4815]: E0225 13:22:11.660031 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:15.660001274 +0000 UTC m=+93.461099338 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 13:22:11 crc kubenswrapper[4815]: E0225 13:22:11.660056 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:15.660042785 +0000 UTC m=+93.461140839 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:11 crc kubenswrapper[4815]: E0225 13:22:11.660088 4815 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 13:22:11 crc kubenswrapper[4815]: E0225 13:22:11.660147 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 13:22:11 crc kubenswrapper[4815]: E0225 13:22:11.660221 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 13:22:11 crc kubenswrapper[4815]: E0225 13:22:11.660231 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:15.66019791 +0000 UTC m=+93.461296024 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 13:22:11 crc kubenswrapper[4815]: E0225 13:22:11.660248 4815 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:11 crc kubenswrapper[4815]: E0225 13:22:11.660399 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:15.660381736 +0000 UTC m=+93.461479820 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.680927 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.680972 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.680983 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.680998 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.681009 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:11Z","lastTransitionTime":"2026-02-25T13:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.783092 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.783140 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.783151 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.783190 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.783221 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:11Z","lastTransitionTime":"2026-02-25T13:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.886695 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.886763 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.886780 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.886803 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.886822 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:11Z","lastTransitionTime":"2026-02-25T13:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.934858 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.934880 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:11 crc kubenswrapper[4815]: E0225 13:22:11.935062 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 13:22:11 crc kubenswrapper[4815]: E0225 13:22:11.935157 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.989591 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.989631 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.989642 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.989659 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:11 crc kubenswrapper[4815]: I0225 13:22:11.989670 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:11Z","lastTransitionTime":"2026-02-25T13:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.092472 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.092528 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.092546 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.092562 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.092573 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:12Z","lastTransitionTime":"2026-02-25T13:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.196279 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.196333 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.196350 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.196371 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.196389 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:12Z","lastTransitionTime":"2026-02-25T13:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.298839 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.298881 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.298894 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.298914 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.298929 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:12Z","lastTransitionTime":"2026-02-25T13:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.401166 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.401209 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.401221 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.401237 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.401247 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:12Z","lastTransitionTime":"2026-02-25T13:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.503304 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.503355 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.503366 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.503390 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.503399 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:12Z","lastTransitionTime":"2026-02-25T13:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.605831 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.605875 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.605891 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.605911 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.605926 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:12Z","lastTransitionTime":"2026-02-25T13:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.708895 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.708955 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.708970 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.708990 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.709009 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:12Z","lastTransitionTime":"2026-02-25T13:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.811116 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.811154 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.811165 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.811179 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.811191 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:12Z","lastTransitionTime":"2026-02-25T13:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.905733 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.905781 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.905798 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.905820 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.905838 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T13:22:12Z","lastTransitionTime":"2026-02-25T13:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 13:22:12 crc kubenswrapper[4815]: I0225 13:22:12.934911 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:12 crc kubenswrapper[4815]: E0225 13:22:12.935342 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 13:22:13 crc kubenswrapper[4815]: I0225 13:22:13.904307 4815 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 25 13:22:13 crc kubenswrapper[4815]: I0225 13:22:13.915651 4815 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 25 13:22:13 crc kubenswrapper[4815]: I0225 13:22:13.935003 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:13 crc kubenswrapper[4815]: I0225 13:22:13.935094 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:13 crc kubenswrapper[4815]: E0225 13:22:13.935560 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 13:22:13 crc kubenswrapper[4815]: E0225 13:22:13.935767 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 13:22:13 crc kubenswrapper[4815]: I0225 13:22:13.949872 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 25 13:22:14 crc kubenswrapper[4815]: I0225 13:22:14.934789 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:14 crc kubenswrapper[4815]: E0225 13:22:14.935858 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 13:22:15 crc kubenswrapper[4815]: I0225 13:22:15.597607 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:15 crc kubenswrapper[4815]: E0225 13:22:15.597811 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:23.597777967 +0000 UTC m=+101.398876031 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:15 crc kubenswrapper[4815]: I0225 13:22:15.698630 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:15 crc kubenswrapper[4815]: I0225 13:22:15.698726 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:15 crc kubenswrapper[4815]: E0225 13:22:15.698757 4815 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 13:22:15 crc kubenswrapper[4815]: E0225 13:22:15.698842 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:23.698820006 +0000 UTC m=+101.499918070 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 13:22:15 crc kubenswrapper[4815]: E0225 13:22:15.698867 4815 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 13:22:15 crc kubenswrapper[4815]: E0225 13:22:15.698887 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 13:22:15 crc kubenswrapper[4815]: E0225 13:22:15.698918 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 13:22:15 crc kubenswrapper[4815]: E0225 13:22:15.698937 4815 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:15 crc kubenswrapper[4815]: E0225 13:22:15.698951 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:23.698927869 +0000 UTC m=+101.500025963 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 13:22:15 crc kubenswrapper[4815]: I0225 13:22:15.698771 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:15 crc kubenswrapper[4815]: E0225 13:22:15.698992 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:23.69897447 +0000 UTC m=+101.500072564 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:15 crc kubenswrapper[4815]: I0225 13:22:15.699022 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:15 crc kubenswrapper[4815]: E0225 13:22:15.699136 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 13:22:15 crc kubenswrapper[4815]: E0225 13:22:15.699157 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 13:22:15 crc kubenswrapper[4815]: E0225 13:22:15.699174 4815 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:15 crc kubenswrapper[4815]: E0225 13:22:15.699252 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:23.699235618 +0000 UTC m=+101.500333712 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:15 crc kubenswrapper[4815]: I0225 13:22:15.964360 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:15 crc kubenswrapper[4815]: I0225 13:22:15.964440 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:15 crc kubenswrapper[4815]: I0225 13:22:15.964389 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:15 crc kubenswrapper[4815]: E0225 13:22:15.964486 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 13:22:15 crc kubenswrapper[4815]: E0225 13:22:15.964583 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 13:22:15 crc kubenswrapper[4815]: E0225 13:22:15.964773 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 13:22:17 crc kubenswrapper[4815]: I0225 13:22:17.934668 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:17 crc kubenswrapper[4815]: I0225 13:22:17.934711 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:17 crc kubenswrapper[4815]: I0225 13:22:17.934772 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:17 crc kubenswrapper[4815]: E0225 13:22:17.934888 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 13:22:17 crc kubenswrapper[4815]: E0225 13:22:17.935075 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 13:22:17 crc kubenswrapper[4815]: E0225 13:22:17.935216 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.496942 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-mbddt"] Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.497607 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-mbddt" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.501086 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.501184 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.501373 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.514858 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-l92cc"] Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.515286 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.515588 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-frc75"] Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.516425 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:22:19 crc kubenswrapper[4815]: W0225 13:22:19.517111 4815 reflector.go:561] object-"openshift-multus"/"default-dockercfg-2q5b6": failed to list *v1.Secret: secrets "default-dockercfg-2q5b6" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Feb 25 13:22:19 crc kubenswrapper[4815]: E0225 13:22:19.517151 4815 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-dockercfg-2q5b6\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-2q5b6\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.517337 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-k8vxf"] Feb 25 13:22:19 crc kubenswrapper[4815]: W0225 13:22:19.517470 4815 reflector.go:561] object-"openshift-multus"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Feb 25 13:22:19 crc kubenswrapper[4815]: E0225 13:22:19.517621 4815 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.518246 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: W0225 13:22:19.519530 4815 reflector.go:561] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": failed to list *v1.Secret: secrets "machine-config-daemon-dockercfg-r5tcq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Feb 25 13:22:19 crc kubenswrapper[4815]: E0225 13:22:19.519575 4815 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"machine-config-daemon-dockercfg-r5tcq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-config-daemon-dockercfg-r5tcq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 25 13:22:19 crc kubenswrapper[4815]: W0225 13:22:19.519715 4815 reflector.go:561] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Feb 25 13:22:19 crc kubenswrapper[4815]: E0225 13:22:19.519747 4815 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 25 13:22:19 crc kubenswrapper[4815]: W0225 13:22:19.519796 4815 reflector.go:561] object-"openshift-machine-config-operator"/"proxy-tls": failed to list *v1.Secret: secrets "proxy-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Feb 25 13:22:19 crc kubenswrapper[4815]: E0225 13:22:19.519815 4815 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"proxy-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"proxy-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 25 13:22:19 crc kubenswrapper[4815]: W0225 13:22:19.519859 4815 reflector.go:561] object-"openshift-multus"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Feb 25 13:22:19 crc kubenswrapper[4815]: E0225 13:22:19.519877 4815 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 25 13:22:19 crc kubenswrapper[4815]: W0225 13:22:19.519976 4815 reflector.go:561] object-"openshift-machine-config-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Feb 25 13:22:19 crc kubenswrapper[4815]: E0225 13:22:19.519997 4815 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 25 13:22:19 crc kubenswrapper[4815]: W0225 13:22:19.520044 4815 reflector.go:561] object-"openshift-machine-config-operator"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Feb 25 13:22:19 crc kubenswrapper[4815]: E0225 13:22:19.520062 4815 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 25 13:22:19 crc kubenswrapper[4815]: W0225 13:22:19.520146 4815 reflector.go:561] object-"openshift-multus"/"multus-daemon-config": failed to list *v1.ConfigMap: configmaps "multus-daemon-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Feb 25 13:22:19 crc kubenswrapper[4815]: E0225 13:22:19.520195 4815 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-daemon-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"multus-daemon-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 25 13:22:19 crc kubenswrapper[4815]: W0225 13:22:19.520499 4815 reflector.go:561] object-"openshift-multus"/"cni-copy-resources": failed to list *v1.ConfigMap: configmaps "cni-copy-resources" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Feb 25 13:22:19 crc kubenswrapper[4815]: E0225 13:22:19.520557 4815 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"cni-copy-resources\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"cni-copy-resources\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.520535 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.520595 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.534888 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qkfgf"] Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.536253 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.538360 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.538562 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.538829 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.539922 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.540149 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.541240 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.541246 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.550303 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=6.550280799 podStartE2EDuration="6.550280799s" podCreationTimestamp="2026-02-25 13:22:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:19.54962996 +0000 UTC m=+97.350728014" watchObservedRunningTime="2026-02-25 13:22:19.550280799 +0000 UTC m=+97.351378853" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.618178 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4"] Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.618530 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.620120 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.620286 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.620397 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.621676 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634007 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-run-multus-certs\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634047 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-cni-bin\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634077 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-cni-binary-copy\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634099 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-var-lib-cni-multus\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634116 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-multus-conf-dir\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634234 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a593fa00-bfb6-44ec-bdbd-441921ae56c5-proxy-tls\") pod \"machine-config-daemon-frc75\" (UID: \"a593fa00-bfb6-44ec-bdbd-441921ae56c5\") " pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634316 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-run-k8s-cni-cncf-io\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634352 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-multus-daemon-config\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634381 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634455 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-var-lib-openvswitch\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634484 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbpb8\" (UniqueName: \"kubernetes.io/projected/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-kube-api-access-cbpb8\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634525 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-slash\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634547 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-log-socket\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634572 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovnkube-config\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634602 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-multus-socket-dir-parent\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634634 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-run-netns\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634659 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/0acc2657-3205-4363-b7a7-d352b615ba75-hosts-file\") pod \"node-resolver-mbddt\" (UID: \"0acc2657-3205-4363-b7a7-d352b615ba75\") " pod="openshift-dns/node-resolver-mbddt" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634684 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-systemd\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634706 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-var-lib-kubelet\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634730 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-etc-kubernetes\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634758 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2c1f35ba-e05e-44ff-bdd4-bf725376b380-cni-binary-copy\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634780 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a593fa00-bfb6-44ec-bdbd-441921ae56c5-mcd-auth-proxy-config\") pod \"machine-config-daemon-frc75\" (UID: \"a593fa00-bfb6-44ec-bdbd-441921ae56c5\") " pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634810 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2c1f35ba-e05e-44ff-bdd4-bf725376b380-system-cni-dir\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634833 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2c1f35ba-e05e-44ff-bdd4-bf725376b380-cnibin\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634854 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2c1f35ba-e05e-44ff-bdd4-bf725376b380-tuning-conf-dir\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634876 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnjqg\" (UniqueName: \"kubernetes.io/projected/0acc2657-3205-4363-b7a7-d352b615ba75-kube-api-access-mnjqg\") pod \"node-resolver-mbddt\" (UID: \"0acc2657-3205-4363-b7a7-d352b615ba75\") " pod="openshift-dns/node-resolver-mbddt" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634904 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-cni-netd\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634929 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-os-release\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634955 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wrw5\" (UniqueName: \"kubernetes.io/projected/a593fa00-bfb6-44ec-bdbd-441921ae56c5-kube-api-access-8wrw5\") pod \"machine-config-daemon-frc75\" (UID: \"a593fa00-bfb6-44ec-bdbd-441921ae56c5\") " pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634977 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-node-log\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.634999 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-run-ovn-kubernetes\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635020 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovnkube-script-lib\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635043 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-hostroot\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635066 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-systemd-units\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635088 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-run-netns\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635113 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-env-overrides\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635138 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blxxb\" (UniqueName: \"kubernetes.io/projected/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-kube-api-access-blxxb\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635160 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-etc-openvswitch\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635179 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-openvswitch\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635200 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-cnibin\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635226 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-var-lib-cni-bin\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635249 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2c1f35ba-e05e-44ff-bdd4-bf725376b380-os-release\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635266 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98lt4\" (UniqueName: \"kubernetes.io/projected/2c1f35ba-e05e-44ff-bdd4-bf725376b380-kube-api-access-98lt4\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635284 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-kubelet\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635308 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-ovn\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635337 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2c1f35ba-e05e-44ff-bdd4-bf725376b380-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635358 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-system-cni-dir\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635379 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-multus-cni-dir\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635402 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a593fa00-bfb6-44ec-bdbd-441921ae56c5-rootfs\") pod \"machine-config-daemon-frc75\" (UID: \"a593fa00-bfb6-44ec-bdbd-441921ae56c5\") " pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.635418 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovn-node-metrics-cert\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.721561 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-zvtx8"] Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.721900 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zvtx8" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.723324 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.723754 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.726960 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.727023 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.735874 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-run-multus-certs\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.735926 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-cni-bin\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.735940 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-run-multus-certs\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.735970 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-cni-binary-copy\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736003 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-var-lib-cni-multus\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736031 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-multus-conf-dir\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736057 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a593fa00-bfb6-44ec-bdbd-441921ae56c5-proxy-tls\") pod \"machine-config-daemon-frc75\" (UID: \"a593fa00-bfb6-44ec-bdbd-441921ae56c5\") " pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736062 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-var-lib-cni-multus\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736056 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-cni-bin\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736096 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-multus-conf-dir\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736114 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-run-k8s-cni-cncf-io\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736081 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-run-k8s-cni-cncf-io\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736185 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-multus-daemon-config\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736232 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736329 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-var-lib-openvswitch\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736334 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736363 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbpb8\" (UniqueName: \"kubernetes.io/projected/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-kube-api-access-cbpb8\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736378 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-var-lib-openvswitch\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736396 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-slash\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736425 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-log-socket\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736454 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovnkube-config\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736495 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-sr8x4\" (UID: \"ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736557 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-sr8x4\" (UID: \"ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736567 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-log-socket\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736593 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-multus-socket-dir-parent\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736657 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-multus-socket-dir-parent\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736700 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-run-netns\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736738 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/0acc2657-3205-4363-b7a7-d352b615ba75-hosts-file\") pod \"node-resolver-mbddt\" (UID: \"0acc2657-3205-4363-b7a7-d352b615ba75\") " pod="openshift-dns/node-resolver-mbddt" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736761 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-systemd\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736790 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-var-lib-kubelet\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736806 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/0acc2657-3205-4363-b7a7-d352b615ba75-hosts-file\") pod \"node-resolver-mbddt\" (UID: \"0acc2657-3205-4363-b7a7-d352b615ba75\") " pod="openshift-dns/node-resolver-mbddt" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736815 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-etc-kubernetes\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736836 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-run-netns\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736866 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-systemd\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736877 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-var-lib-kubelet\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736847 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2c1f35ba-e05e-44ff-bdd4-bf725376b380-cni-binary-copy\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736924 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-etc-kubernetes\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736925 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-sr8x4\" (UID: \"ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.736971 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a593fa00-bfb6-44ec-bdbd-441921ae56c5-mcd-auth-proxy-config\") pod \"machine-config-daemon-frc75\" (UID: \"a593fa00-bfb6-44ec-bdbd-441921ae56c5\") " pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737001 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2c1f35ba-e05e-44ff-bdd4-bf725376b380-system-cni-dir\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737036 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2c1f35ba-e05e-44ff-bdd4-bf725376b380-cnibin\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737058 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2c1f35ba-e05e-44ff-bdd4-bf725376b380-tuning-conf-dir\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737080 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnjqg\" (UniqueName: \"kubernetes.io/projected/0acc2657-3205-4363-b7a7-d352b615ba75-kube-api-access-mnjqg\") pod \"node-resolver-mbddt\" (UID: \"0acc2657-3205-4363-b7a7-d352b615ba75\") " pod="openshift-dns/node-resolver-mbddt" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737109 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-cni-netd\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737124 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2c1f35ba-e05e-44ff-bdd4-bf725376b380-cnibin\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737134 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-os-release\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737160 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovnkube-config\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737184 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wrw5\" (UniqueName: \"kubernetes.io/projected/a593fa00-bfb6-44ec-bdbd-441921ae56c5-kube-api-access-8wrw5\") pod \"machine-config-daemon-frc75\" (UID: \"a593fa00-bfb6-44ec-bdbd-441921ae56c5\") " pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737214 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-cni-netd\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737220 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-node-log\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737252 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-run-ovn-kubernetes\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737287 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovnkube-script-lib\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737322 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-hostroot\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737353 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-systemd-units\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737377 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-run-netns\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737107 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2c1f35ba-e05e-44ff-bdd4-bf725376b380-system-cni-dir\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737407 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-env-overrides\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737433 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-node-log\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737222 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-os-release\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737447 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blxxb\" (UniqueName: \"kubernetes.io/projected/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-kube-api-access-blxxb\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737496 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-etc-openvswitch\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737550 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-openvswitch\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737573 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a-service-ca\") pod \"cluster-version-operator-5c965bbfc6-sr8x4\" (UID: \"ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737595 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-cnibin\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737616 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-var-lib-cni-bin\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737641 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2c1f35ba-e05e-44ff-bdd4-bf725376b380-os-release\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737662 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98lt4\" (UniqueName: \"kubernetes.io/projected/2c1f35ba-e05e-44ff-bdd4-bf725376b380-kube-api-access-98lt4\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737674 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-openvswitch\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737682 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-kubelet\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737710 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-kubelet\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737738 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-ovn\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737768 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-cnibin\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737800 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-host-var-lib-cni-bin\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737801 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2c1f35ba-e05e-44ff-bdd4-bf725376b380-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737840 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-system-cni-dir\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737849 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2c1f35ba-e05e-44ff-bdd4-bf725376b380-os-release\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737857 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2c1f35ba-e05e-44ff-bdd4-bf725376b380-tuning-conf-dir\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737877 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-multus-cni-dir\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737896 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-systemd-units\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737920 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a593fa00-bfb6-44ec-bdbd-441921ae56c5-rootfs\") pod \"machine-config-daemon-frc75\" (UID: \"a593fa00-bfb6-44ec-bdbd-441921ae56c5\") " pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.737968 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-hostroot\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.738001 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovn-node-metrics-cert\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.738023 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a593fa00-bfb6-44ec-bdbd-441921ae56c5-rootfs\") pod \"machine-config-daemon-frc75\" (UID: \"a593fa00-bfb6-44ec-bdbd-441921ae56c5\") " pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.738026 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-sr8x4\" (UID: \"ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.738114 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-slash\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.738180 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-run-ovn-kubernetes\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.738237 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-etc-openvswitch\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.738280 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-ovn\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.738667 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-env-overrides\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.738672 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovnkube-script-lib\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.738727 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-system-cni-dir\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.738741 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-run-netns\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.738757 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-multus-cni-dir\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.739314 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2c1f35ba-e05e-44ff-bdd4-bf725376b380-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.754442 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovn-node-metrics-cert\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.756731 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnjqg\" (UniqueName: \"kubernetes.io/projected/0acc2657-3205-4363-b7a7-d352b615ba75-kube-api-access-mnjqg\") pod \"node-resolver-mbddt\" (UID: \"0acc2657-3205-4363-b7a7-d352b615ba75\") " pod="openshift-dns/node-resolver-mbddt" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.761107 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbpb8\" (UniqueName: \"kubernetes.io/projected/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-kube-api-access-cbpb8\") pod \"ovnkube-node-qkfgf\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.812009 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-mbddt" Feb 25 13:22:19 crc kubenswrapper[4815]: W0225 13:22:19.824459 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0acc2657_3205_4363_b7a7_d352b615ba75.slice/crio-b16a63c3784e58a600e5c651f47fb2cc84f7440a7563be6721c9ac250d0e7afe WatchSource:0}: Error finding container b16a63c3784e58a600e5c651f47fb2cc84f7440a7563be6721c9ac250d0e7afe: Status 404 returned error can't find the container with id b16a63c3784e58a600e5c651f47fb2cc84f7440a7563be6721c9ac250d0e7afe Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.838797 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-sr8x4\" (UID: \"ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.838838 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/0419ad2d-25ed-4238-bf34-248566e88632-serviceca\") pod \"node-ca-zvtx8\" (UID: \"0419ad2d-25ed-4238-bf34-248566e88632\") " pod="openshift-image-registry/node-ca-zvtx8" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.838894 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-sr8x4\" (UID: \"ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.838912 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-sr8x4\" (UID: \"ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.838929 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-sr8x4\" (UID: \"ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.838925 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-sr8x4\" (UID: \"ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.839006 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndd77\" (UniqueName: \"kubernetes.io/projected/0419ad2d-25ed-4238-bf34-248566e88632-kube-api-access-ndd77\") pod \"node-ca-zvtx8\" (UID: \"0419ad2d-25ed-4238-bf34-248566e88632\") " pod="openshift-image-registry/node-ca-zvtx8" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.839050 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a-service-ca\") pod \"cluster-version-operator-5c965bbfc6-sr8x4\" (UID: \"ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.839078 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0419ad2d-25ed-4238-bf34-248566e88632-host\") pod \"node-ca-zvtx8\" (UID: \"0419ad2d-25ed-4238-bf34-248566e88632\") " pod="openshift-image-registry/node-ca-zvtx8" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.839191 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-sr8x4\" (UID: \"ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.840064 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a-service-ca\") pod \"cluster-version-operator-5c965bbfc6-sr8x4\" (UID: \"ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.843193 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-sr8x4\" (UID: \"ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.856815 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.861868 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-sr8x4\" (UID: \"ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" Feb 25 13:22:19 crc kubenswrapper[4815]: W0225 13:22:19.872806 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6cd6cca6_d1e9_4c56_8f4d_f396a769ba07.slice/crio-0d7781213d2e22dc8880868392952084c00d31ea6d52ec3244161ae3b53ca60e WatchSource:0}: Error finding container 0d7781213d2e22dc8880868392952084c00d31ea6d52ec3244161ae3b53ca60e: Status 404 returned error can't find the container with id 0d7781213d2e22dc8880868392952084c00d31ea6d52ec3244161ae3b53ca60e Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.874109 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp"] Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.874491 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.876260 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.876530 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.891977 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-gbqbs"] Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.892374 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:19 crc kubenswrapper[4815]: E0225 13:22:19.892441 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbqbs" podUID="97b044d0-af42-4f42-95a8-49d6c8782246" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.929033 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.934673 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.934724 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.934741 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:19 crc kubenswrapper[4815]: E0225 13:22:19.934866 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 13:22:19 crc kubenswrapper[4815]: E0225 13:22:19.935098 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 13:22:19 crc kubenswrapper[4815]: E0225 13:22:19.935214 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 13:22:19 crc kubenswrapper[4815]: W0225 13:22:19.939226 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec7fe59f_3ca4_40bb_b87e_f65dcb1f302a.slice/crio-55fa2b839bcb6200cd3fe68fae14c5f1aa9b864b37e63e2cfb490cb0195a0dd8 WatchSource:0}: Error finding container 55fa2b839bcb6200cd3fe68fae14c5f1aa9b864b37e63e2cfb490cb0195a0dd8: Status 404 returned error can't find the container with id 55fa2b839bcb6200cd3fe68fae14c5f1aa9b864b37e63e2cfb490cb0195a0dd8 Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.939646 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlgg5\" (UniqueName: \"kubernetes.io/projected/88e70f83-2c7e-46ce-896b-94d5672a9efd-kube-api-access-qlgg5\") pod \"ovnkube-control-plane-749d76644c-pwvdp\" (UID: \"88e70f83-2c7e-46ce-896b-94d5672a9efd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.939699 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/88e70f83-2c7e-46ce-896b-94d5672a9efd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pwvdp\" (UID: \"88e70f83-2c7e-46ce-896b-94d5672a9efd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.939914 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndd77\" (UniqueName: \"kubernetes.io/projected/0419ad2d-25ed-4238-bf34-248566e88632-kube-api-access-ndd77\") pod \"node-ca-zvtx8\" (UID: \"0419ad2d-25ed-4238-bf34-248566e88632\") " pod="openshift-image-registry/node-ca-zvtx8" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.939991 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0419ad2d-25ed-4238-bf34-248566e88632-host\") pod \"node-ca-zvtx8\" (UID: \"0419ad2d-25ed-4238-bf34-248566e88632\") " pod="openshift-image-registry/node-ca-zvtx8" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.940053 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/88e70f83-2c7e-46ce-896b-94d5672a9efd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pwvdp\" (UID: \"88e70f83-2c7e-46ce-896b-94d5672a9efd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.940092 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/0419ad2d-25ed-4238-bf34-248566e88632-serviceca\") pod \"node-ca-zvtx8\" (UID: \"0419ad2d-25ed-4238-bf34-248566e88632\") " pod="openshift-image-registry/node-ca-zvtx8" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.940098 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0419ad2d-25ed-4238-bf34-248566e88632-host\") pod \"node-ca-zvtx8\" (UID: \"0419ad2d-25ed-4238-bf34-248566e88632\") " pod="openshift-image-registry/node-ca-zvtx8" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.940120 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff7mn\" (UniqueName: \"kubernetes.io/projected/97b044d0-af42-4f42-95a8-49d6c8782246-kube-api-access-ff7mn\") pod \"network-metrics-daemon-gbqbs\" (UID: \"97b044d0-af42-4f42-95a8-49d6c8782246\") " pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.940203 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/88e70f83-2c7e-46ce-896b-94d5672a9efd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pwvdp\" (UID: \"88e70f83-2c7e-46ce-896b-94d5672a9efd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.940256 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs\") pod \"network-metrics-daemon-gbqbs\" (UID: \"97b044d0-af42-4f42-95a8-49d6c8782246\") " pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.940984 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/0419ad2d-25ed-4238-bf34-248566e88632-serviceca\") pod \"node-ca-zvtx8\" (UID: \"0419ad2d-25ed-4238-bf34-248566e88632\") " pod="openshift-image-registry/node-ca-zvtx8" Feb 25 13:22:19 crc kubenswrapper[4815]: I0225 13:22:19.956455 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndd77\" (UniqueName: \"kubernetes.io/projected/0419ad2d-25ed-4238-bf34-248566e88632-kube-api-access-ndd77\") pod \"node-ca-zvtx8\" (UID: \"0419ad2d-25ed-4238-bf34-248566e88632\") " pod="openshift-image-registry/node-ca-zvtx8" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.036433 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zvtx8" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.040910 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/88e70f83-2c7e-46ce-896b-94d5672a9efd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pwvdp\" (UID: \"88e70f83-2c7e-46ce-896b-94d5672a9efd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.040977 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs\") pod \"network-metrics-daemon-gbqbs\" (UID: \"97b044d0-af42-4f42-95a8-49d6c8782246\") " pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.041016 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlgg5\" (UniqueName: \"kubernetes.io/projected/88e70f83-2c7e-46ce-896b-94d5672a9efd-kube-api-access-qlgg5\") pod \"ovnkube-control-plane-749d76644c-pwvdp\" (UID: \"88e70f83-2c7e-46ce-896b-94d5672a9efd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.041080 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/88e70f83-2c7e-46ce-896b-94d5672a9efd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pwvdp\" (UID: \"88e70f83-2c7e-46ce-896b-94d5672a9efd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.041174 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/88e70f83-2c7e-46ce-896b-94d5672a9efd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pwvdp\" (UID: \"88e70f83-2c7e-46ce-896b-94d5672a9efd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" Feb 25 13:22:20 crc kubenswrapper[4815]: E0225 13:22:20.041189 4815 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.041219 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff7mn\" (UniqueName: \"kubernetes.io/projected/97b044d0-af42-4f42-95a8-49d6c8782246-kube-api-access-ff7mn\") pod \"network-metrics-daemon-gbqbs\" (UID: \"97b044d0-af42-4f42-95a8-49d6c8782246\") " pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:20 crc kubenswrapper[4815]: E0225 13:22:20.041265 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs podName:97b044d0-af42-4f42-95a8-49d6c8782246 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:20.541244268 +0000 UTC m=+98.342342312 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs") pod "network-metrics-daemon-gbqbs" (UID: "97b044d0-af42-4f42-95a8-49d6c8782246") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.041659 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/88e70f83-2c7e-46ce-896b-94d5672a9efd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pwvdp\" (UID: \"88e70f83-2c7e-46ce-896b-94d5672a9efd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.041897 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/88e70f83-2c7e-46ce-896b-94d5672a9efd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pwvdp\" (UID: \"88e70f83-2c7e-46ce-896b-94d5672a9efd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.044282 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/88e70f83-2c7e-46ce-896b-94d5672a9efd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pwvdp\" (UID: \"88e70f83-2c7e-46ce-896b-94d5672a9efd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" Feb 25 13:22:20 crc kubenswrapper[4815]: W0225 13:22:20.047619 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0419ad2d_25ed_4238_bf34_248566e88632.slice/crio-b271599964c0c0f22b6bdbbc275dc88ea3850539cf43917040bf9f83597af06d WatchSource:0}: Error finding container b271599964c0c0f22b6bdbbc275dc88ea3850539cf43917040bf9f83597af06d: Status 404 returned error can't find the container with id b271599964c0c0f22b6bdbbc275dc88ea3850539cf43917040bf9f83597af06d Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.060601 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlgg5\" (UniqueName: \"kubernetes.io/projected/88e70f83-2c7e-46ce-896b-94d5672a9efd-kube-api-access-qlgg5\") pod \"ovnkube-control-plane-749d76644c-pwvdp\" (UID: \"88e70f83-2c7e-46ce-896b-94d5672a9efd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.193004 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" Feb 25 13:22:20 crc kubenswrapper[4815]: W0225 13:22:20.208599 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88e70f83_2c7e_46ce_896b_94d5672a9efd.slice/crio-d49c7bf4819ba9078eaf78f2c18269ac4ac492f8e895bbe60b21eba0c5857af9 WatchSource:0}: Error finding container d49c7bf4819ba9078eaf78f2c18269ac4ac492f8e895bbe60b21eba0c5857af9: Status 404 returned error can't find the container with id d49c7bf4819ba9078eaf78f2c18269ac4ac492f8e895bbe60b21eba0c5857af9 Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.406827 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zvtx8" event={"ID":"0419ad2d-25ed-4238-bf34-248566e88632","Type":"ContainerStarted","Data":"a658a422cf0830cfccdb8158afe9adbe5dbe79c4a1e188b0b5279f373c4d3550"} Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.406878 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zvtx8" event={"ID":"0419ad2d-25ed-4238-bf34-248566e88632","Type":"ContainerStarted","Data":"b271599964c0c0f22b6bdbbc275dc88ea3850539cf43917040bf9f83597af06d"} Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.409376 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" event={"ID":"ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a","Type":"ContainerStarted","Data":"d974b8e6797a0e79fad84e807f994a1073d61ecf504f9467faede7a8209e6014"} Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.409429 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" event={"ID":"ec7fe59f-3ca4-40bb-b87e-f65dcb1f302a","Type":"ContainerStarted","Data":"55fa2b839bcb6200cd3fe68fae14c5f1aa9b864b37e63e2cfb490cb0195a0dd8"} Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.413238 4815 generic.go:334] "Generic (PLEG): container finished" podID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerID="ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085" exitCode=0 Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.413326 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerDied","Data":"ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085"} Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.413470 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerStarted","Data":"0d7781213d2e22dc8880868392952084c00d31ea6d52ec3244161ae3b53ca60e"} Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.415333 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" event={"ID":"88e70f83-2c7e-46ce-896b-94d5672a9efd","Type":"ContainerStarted","Data":"4dbbbff7e3c34ef2ef76d1b52dd9cd48f0c3a23920e4584a9e2b5454c198ee6a"} Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.415391 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" event={"ID":"88e70f83-2c7e-46ce-896b-94d5672a9efd","Type":"ContainerStarted","Data":"d49c7bf4819ba9078eaf78f2c18269ac4ac492f8e895bbe60b21eba0c5857af9"} Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.419699 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-mbddt" event={"ID":"0acc2657-3205-4363-b7a7-d352b615ba75","Type":"ContainerStarted","Data":"91f0cdb04294f516b62325d62bf699fe08adb269ba946f7335846dc56a309d55"} Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.419812 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-mbddt" event={"ID":"0acc2657-3205-4363-b7a7-d352b615ba75","Type":"ContainerStarted","Data":"b16a63c3784e58a600e5c651f47fb2cc84f7440a7563be6721c9ac250d0e7afe"} Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.429545 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-sr8x4" podStartSLOduration=45.429500618 podStartE2EDuration="45.429500618s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:20.427819327 +0000 UTC m=+98.228917391" watchObservedRunningTime="2026-02-25 13:22:20.429500618 +0000 UTC m=+98.230598672" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.476696 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.478841 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-mbddt" podStartSLOduration=45.478824486 podStartE2EDuration="45.478824486s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:20.477827385 +0000 UTC m=+98.278925459" watchObservedRunningTime="2026-02-25 13:22:20.478824486 +0000 UTC m=+98.279922540" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.546717 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs\") pod \"network-metrics-daemon-gbqbs\" (UID: \"97b044d0-af42-4f42-95a8-49d6c8782246\") " pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:20 crc kubenswrapper[4815]: E0225 13:22:20.547625 4815 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 13:22:20 crc kubenswrapper[4815]: E0225 13:22:20.547704 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs podName:97b044d0-af42-4f42-95a8-49d6c8782246 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:21.547681106 +0000 UTC m=+99.348779250 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs") pod "network-metrics-daemon-gbqbs" (UID: "97b044d0-af42-4f42-95a8-49d6c8782246") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.617006 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.618001 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2c1f35ba-e05e-44ff-bdd4-bf725376b380-cni-binary-copy\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.624769 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.627745 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-cni-binary-copy\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.647947 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.678195 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.687428 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-multus-daemon-config\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:20 crc kubenswrapper[4815]: E0225 13:22:20.736452 4815 secret.go:188] Couldn't get secret openshift-machine-config-operator/proxy-tls: failed to sync secret cache: timed out waiting for the condition Feb 25 13:22:20 crc kubenswrapper[4815]: E0225 13:22:20.736541 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a593fa00-bfb6-44ec-bdbd-441921ae56c5-proxy-tls podName:a593fa00-bfb6-44ec-bdbd-441921ae56c5 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:21.236518101 +0000 UTC m=+99.037616155 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/a593fa00-bfb6-44ec-bdbd-441921ae56c5-proxy-tls") pod "machine-config-daemon-frc75" (UID: "a593fa00-bfb6-44ec-bdbd-441921ae56c5") : failed to sync secret cache: timed out waiting for the condition Feb 25 13:22:20 crc kubenswrapper[4815]: E0225 13:22:20.739630 4815 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Feb 25 13:22:20 crc kubenswrapper[4815]: E0225 13:22:20.739748 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a593fa00-bfb6-44ec-bdbd-441921ae56c5-mcd-auth-proxy-config podName:a593fa00-bfb6-44ec-bdbd-441921ae56c5 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:21.239728599 +0000 UTC m=+99.040826653 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "mcd-auth-proxy-config" (UniqueName: "kubernetes.io/configmap/a593fa00-bfb6-44ec-bdbd-441921ae56c5-mcd-auth-proxy-config") pod "machine-config-daemon-frc75" (UID: "a593fa00-bfb6-44ec-bdbd-441921ae56c5") : failed to sync configmap cache: timed out waiting for the condition Feb 25 13:22:20 crc kubenswrapper[4815]: E0225 13:22:20.754256 4815 projected.go:288] Couldn't get configMap openshift-multus/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Feb 25 13:22:20 crc kubenswrapper[4815]: E0225 13:22:20.754583 4815 projected.go:194] Error preparing data for projected volume kube-api-access-98lt4 for pod openshift-multus/multus-additional-cni-plugins-k8vxf: failed to sync configmap cache: timed out waiting for the condition Feb 25 13:22:20 crc kubenswrapper[4815]: E0225 13:22:20.754634 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2c1f35ba-e05e-44ff-bdd4-bf725376b380-kube-api-access-98lt4 podName:2c1f35ba-e05e-44ff-bdd4-bf725376b380 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:21.25461651 +0000 UTC m=+99.055714564 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-98lt4" (UniqueName: "kubernetes.io/projected/2c1f35ba-e05e-44ff-bdd4-bf725376b380-kube-api-access-98lt4") pod "multus-additional-cni-plugins-k8vxf" (UID: "2c1f35ba-e05e-44ff-bdd4-bf725376b380") : failed to sync configmap cache: timed out waiting for the condition Feb 25 13:22:20 crc kubenswrapper[4815]: E0225 13:22:20.754284 4815 projected.go:288] Couldn't get configMap openshift-machine-config-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Feb 25 13:22:20 crc kubenswrapper[4815]: E0225 13:22:20.754661 4815 projected.go:194] Error preparing data for projected volume kube-api-access-8wrw5 for pod openshift-machine-config-operator/machine-config-daemon-frc75: failed to sync configmap cache: timed out waiting for the condition Feb 25 13:22:20 crc kubenswrapper[4815]: E0225 13:22:20.754686 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a593fa00-bfb6-44ec-bdbd-441921ae56c5-kube-api-access-8wrw5 podName:a593fa00-bfb6-44ec-bdbd-441921ae56c5 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:21.254678262 +0000 UTC m=+99.055776316 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-8wrw5" (UniqueName: "kubernetes.io/projected/a593fa00-bfb6-44ec-bdbd-441921ae56c5-kube-api-access-8wrw5") pod "machine-config-daemon-frc75" (UID: "a593fa00-bfb6-44ec-bdbd-441921ae56c5") : failed to sync configmap cache: timed out waiting for the condition Feb 25 13:22:20 crc kubenswrapper[4815]: E0225 13:22:20.769243 4815 projected.go:288] Couldn't get configMap openshift-multus/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Feb 25 13:22:20 crc kubenswrapper[4815]: E0225 13:22:20.769317 4815 projected.go:194] Error preparing data for projected volume kube-api-access-blxxb for pod openshift-multus/multus-l92cc: failed to sync configmap cache: timed out waiting for the condition Feb 25 13:22:20 crc kubenswrapper[4815]: E0225 13:22:20.769379 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-kube-api-access-blxxb podName:88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:21.269361198 +0000 UTC m=+99.070459252 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-blxxb" (UniqueName: "kubernetes.io/projected/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-kube-api-access-blxxb") pod "multus-l92cc" (UID: "88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6") : failed to sync configmap cache: timed out waiting for the condition Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.824262 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.834300 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff7mn\" (UniqueName: \"kubernetes.io/projected/97b044d0-af42-4f42-95a8-49d6c8782246-kube-api-access-ff7mn\") pod \"network-metrics-daemon-gbqbs\" (UID: \"97b044d0-af42-4f42-95a8-49d6c8782246\") " pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:20 crc kubenswrapper[4815]: I0225 13:22:20.981744 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.076230 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.123693 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.124143 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.256952 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a593fa00-bfb6-44ec-bdbd-441921ae56c5-proxy-tls\") pod \"machine-config-daemon-frc75\" (UID: \"a593fa00-bfb6-44ec-bdbd-441921ae56c5\") " pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.257016 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a593fa00-bfb6-44ec-bdbd-441921ae56c5-mcd-auth-proxy-config\") pod \"machine-config-daemon-frc75\" (UID: \"a593fa00-bfb6-44ec-bdbd-441921ae56c5\") " pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.257035 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wrw5\" (UniqueName: \"kubernetes.io/projected/a593fa00-bfb6-44ec-bdbd-441921ae56c5-kube-api-access-8wrw5\") pod \"machine-config-daemon-frc75\" (UID: \"a593fa00-bfb6-44ec-bdbd-441921ae56c5\") " pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.257060 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98lt4\" (UniqueName: \"kubernetes.io/projected/2c1f35ba-e05e-44ff-bdd4-bf725376b380-kube-api-access-98lt4\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.258203 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a593fa00-bfb6-44ec-bdbd-441921ae56c5-mcd-auth-proxy-config\") pod \"machine-config-daemon-frc75\" (UID: \"a593fa00-bfb6-44ec-bdbd-441921ae56c5\") " pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.260185 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98lt4\" (UniqueName: \"kubernetes.io/projected/2c1f35ba-e05e-44ff-bdd4-bf725376b380-kube-api-access-98lt4\") pod \"multus-additional-cni-plugins-k8vxf\" (UID: \"2c1f35ba-e05e-44ff-bdd4-bf725376b380\") " pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.260618 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a593fa00-bfb6-44ec-bdbd-441921ae56c5-proxy-tls\") pod \"machine-config-daemon-frc75\" (UID: \"a593fa00-bfb6-44ec-bdbd-441921ae56c5\") " pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.262812 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wrw5\" (UniqueName: \"kubernetes.io/projected/a593fa00-bfb6-44ec-bdbd-441921ae56c5-kube-api-access-8wrw5\") pod \"machine-config-daemon-frc75\" (UID: \"a593fa00-bfb6-44ec-bdbd-441921ae56c5\") " pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.340738 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:22:21 crc kubenswrapper[4815]: W0225 13:22:21.351693 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda593fa00_bfb6_44ec_bdbd_441921ae56c5.slice/crio-c489a1de867cbd6c46a768ceb8182a76294fa11648022fb40413fb581d84bb90 WatchSource:0}: Error finding container c489a1de867cbd6c46a768ceb8182a76294fa11648022fb40413fb581d84bb90: Status 404 returned error can't find the container with id c489a1de867cbd6c46a768ceb8182a76294fa11648022fb40413fb581d84bb90 Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.352543 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-k8vxf" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.357970 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blxxb\" (UniqueName: \"kubernetes.io/projected/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-kube-api-access-blxxb\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.362014 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blxxb\" (UniqueName: \"kubernetes.io/projected/88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6-kube-api-access-blxxb\") pod \"multus-l92cc\" (UID: \"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6\") " pod="openshift-multus/multus-l92cc" Feb 25 13:22:21 crc kubenswrapper[4815]: W0225 13:22:21.372863 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c1f35ba_e05e_44ff_bdd4_bf725376b380.slice/crio-6b343014b25c080bf2d9ef2a88d02bbb5b2f6d6107b859e860124eb09c64713e WatchSource:0}: Error finding container 6b343014b25c080bf2d9ef2a88d02bbb5b2f6d6107b859e860124eb09c64713e: Status 404 returned error can't find the container with id 6b343014b25c080bf2d9ef2a88d02bbb5b2f6d6107b859e860124eb09c64713e Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.424694 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-k8vxf" event={"ID":"2c1f35ba-e05e-44ff-bdd4-bf725376b380","Type":"ContainerStarted","Data":"6b343014b25c080bf2d9ef2a88d02bbb5b2f6d6107b859e860124eb09c64713e"} Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.434727 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" event={"ID":"88e70f83-2c7e-46ce-896b-94d5672a9efd","Type":"ContainerStarted","Data":"0558e3456887546d57d97f071dd5855cc1a8ff10bacf283f58b6a4e5094b91a8"} Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.443212 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerStarted","Data":"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285"} Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.443256 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerStarted","Data":"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0"} Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.443270 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerStarted","Data":"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820"} Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.443284 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerStarted","Data":"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4"} Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.443296 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerStarted","Data":"9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743"} Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.443305 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerStarted","Data":"8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90"} Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.444833 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerStarted","Data":"c489a1de867cbd6c46a768ceb8182a76294fa11648022fb40413fb581d84bb90"} Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.453726 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwvdp" podStartSLOduration=45.45371141 podStartE2EDuration="45.45371141s" podCreationTimestamp="2026-02-25 13:21:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:21.452584815 +0000 UTC m=+99.253682899" watchObservedRunningTime="2026-02-25 13:22:21.45371141 +0000 UTC m=+99.254809464" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.465550 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-zvtx8" podStartSLOduration=46.465528218 podStartE2EDuration="46.465528218s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:21.465035203 +0000 UTC m=+99.266133267" watchObservedRunningTime="2026-02-25 13:22:21.465528218 +0000 UTC m=+99.266626282" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.560928 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs\") pod \"network-metrics-daemon-gbqbs\" (UID: \"97b044d0-af42-4f42-95a8-49d6c8782246\") " pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:21 crc kubenswrapper[4815]: E0225 13:22:21.562530 4815 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 13:22:21 crc kubenswrapper[4815]: E0225 13:22:21.562592 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs podName:97b044d0-af42-4f42-95a8-49d6c8782246 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:23.562576565 +0000 UTC m=+101.363674609 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs") pod "network-metrics-daemon-gbqbs" (UID: "97b044d0-af42-4f42-95a8-49d6c8782246") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.631439 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-l92cc" Feb 25 13:22:21 crc kubenswrapper[4815]: W0225 13:22:21.644107 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88b40c3e_14f9_4ec2_bd13_2fecf33ec9e6.slice/crio-0fd95f61b12a9d55471c554b7b367a812749c62ffff57866684addda445f0e17 WatchSource:0}: Error finding container 0fd95f61b12a9d55471c554b7b367a812749c62ffff57866684addda445f0e17: Status 404 returned error can't find the container with id 0fd95f61b12a9d55471c554b7b367a812749c62ffff57866684addda445f0e17 Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.935593 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.935593 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.935772 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:21 crc kubenswrapper[4815]: E0225 13:22:21.935892 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 13:22:21 crc kubenswrapper[4815]: E0225 13:22:21.935954 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 13:22:21 crc kubenswrapper[4815]: E0225 13:22:21.936121 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbqbs" podUID="97b044d0-af42-4f42-95a8-49d6c8782246" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.936211 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:21 crc kubenswrapper[4815]: E0225 13:22:21.936310 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 13:22:21 crc kubenswrapper[4815]: I0225 13:22:21.951884 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 25 13:22:22 crc kubenswrapper[4815]: I0225 13:22:22.449055 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-l92cc" event={"ID":"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6","Type":"ContainerStarted","Data":"86c30dd3c376e2edfbfc035c5f37c6d85c659e8fe429f8c162ca3e14df9048c5"} Feb 25 13:22:22 crc kubenswrapper[4815]: I0225 13:22:22.449392 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-l92cc" event={"ID":"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6","Type":"ContainerStarted","Data":"0fd95f61b12a9d55471c554b7b367a812749c62ffff57866684addda445f0e17"} Feb 25 13:22:22 crc kubenswrapper[4815]: I0225 13:22:22.451300 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerStarted","Data":"af1f287c5a99fd8cc7e8326d2fe180a911200a07983a4f07f7eae7272d0b1b6b"} Feb 25 13:22:22 crc kubenswrapper[4815]: I0225 13:22:22.451328 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerStarted","Data":"440b745816f3d3542a46a14793d9d030a841eaa34e91cd17885d1961ecc85bee"} Feb 25 13:22:22 crc kubenswrapper[4815]: I0225 13:22:22.453606 4815 generic.go:334] "Generic (PLEG): container finished" podID="2c1f35ba-e05e-44ff-bdd4-bf725376b380" containerID="6b7791e4a8bee7a820b7bfbba33c26dc98f7599427382fb080c2988697bfb97d" exitCode=0 Feb 25 13:22:22 crc kubenswrapper[4815]: I0225 13:22:22.453735 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-k8vxf" event={"ID":"2c1f35ba-e05e-44ff-bdd4-bf725376b380","Type":"ContainerDied","Data":"6b7791e4a8bee7a820b7bfbba33c26dc98f7599427382fb080c2988697bfb97d"} Feb 25 13:22:22 crc kubenswrapper[4815]: I0225 13:22:22.478937 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.47891427 podStartE2EDuration="1.47891427s" podCreationTimestamp="2026-02-25 13:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:22.477408315 +0000 UTC m=+100.278506399" watchObservedRunningTime="2026-02-25 13:22:22.47891427 +0000 UTC m=+100.280012344" Feb 25 13:22:22 crc kubenswrapper[4815]: I0225 13:22:22.515277 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-l92cc" podStartSLOduration=47.515255444 podStartE2EDuration="47.515255444s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:22.498291688 +0000 UTC m=+100.299389782" watchObservedRunningTime="2026-02-25 13:22:22.515255444 +0000 UTC m=+100.316353498" Feb 25 13:22:22 crc kubenswrapper[4815]: I0225 13:22:22.515536 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podStartSLOduration=47.515502181 podStartE2EDuration="47.515502181s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:22.514715428 +0000 UTC m=+100.315813522" watchObservedRunningTime="2026-02-25 13:22:22.515502181 +0000 UTC m=+100.316600235" Feb 25 13:22:23 crc kubenswrapper[4815]: I0225 13:22:23.463675 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerStarted","Data":"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4"} Feb 25 13:22:23 crc kubenswrapper[4815]: I0225 13:22:23.465948 4815 generic.go:334] "Generic (PLEG): container finished" podID="2c1f35ba-e05e-44ff-bdd4-bf725376b380" containerID="7d0a2560f8747835dd42e94a504b341e5596ca666c3b8eedb3fb1c40ad546269" exitCode=0 Feb 25 13:22:23 crc kubenswrapper[4815]: I0225 13:22:23.466007 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-k8vxf" event={"ID":"2c1f35ba-e05e-44ff-bdd4-bf725376b380","Type":"ContainerDied","Data":"7d0a2560f8747835dd42e94a504b341e5596ca666c3b8eedb3fb1c40ad546269"} Feb 25 13:22:23 crc kubenswrapper[4815]: I0225 13:22:23.585450 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs\") pod \"network-metrics-daemon-gbqbs\" (UID: \"97b044d0-af42-4f42-95a8-49d6c8782246\") " pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.586213 4815 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.586275 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs podName:97b044d0-af42-4f42-95a8-49d6c8782246 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:27.586257116 +0000 UTC m=+105.387355260 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs") pod "network-metrics-daemon-gbqbs" (UID: "97b044d0-af42-4f42-95a8-49d6c8782246") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 13:22:23 crc kubenswrapper[4815]: I0225 13:22:23.686123 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.686306 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:39.686290174 +0000 UTC m=+117.487388228 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:23 crc kubenswrapper[4815]: I0225 13:22:23.786863 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:23 crc kubenswrapper[4815]: I0225 13:22:23.787129 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:23 crc kubenswrapper[4815]: I0225 13:22:23.787149 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.787079 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.787222 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.787235 4815 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.787279 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.787290 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:39.78727367 +0000 UTC m=+117.588371724 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.787294 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.787298 4815 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.787326 4815 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.787309 4815 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.787365 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:39.787352622 +0000 UTC m=+117.588450676 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 13:22:23 crc kubenswrapper[4815]: I0225 13:22:23.787189 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.787378 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:39.787373613 +0000 UTC m=+117.588471667 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.787388 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:39.787383753 +0000 UTC m=+117.588481807 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 13:22:23 crc kubenswrapper[4815]: I0225 13:22:23.934557 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:23 crc kubenswrapper[4815]: I0225 13:22:23.934638 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:23 crc kubenswrapper[4815]: I0225 13:22:23.934676 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.934768 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.934684 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 13:22:23 crc kubenswrapper[4815]: I0225 13:22:23.934710 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.934919 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbqbs" podUID="97b044d0-af42-4f42-95a8-49d6c8782246" Feb 25 13:22:23 crc kubenswrapper[4815]: E0225 13:22:23.935112 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 13:22:24 crc kubenswrapper[4815]: I0225 13:22:24.471329 4815 generic.go:334] "Generic (PLEG): container finished" podID="2c1f35ba-e05e-44ff-bdd4-bf725376b380" containerID="b429fe042904202ddcb897e33e35bfe5e5f9ec1d745e6eaf631b2c3303badfd8" exitCode=0 Feb 25 13:22:24 crc kubenswrapper[4815]: I0225 13:22:24.471384 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-k8vxf" event={"ID":"2c1f35ba-e05e-44ff-bdd4-bf725376b380","Type":"ContainerDied","Data":"b429fe042904202ddcb897e33e35bfe5e5f9ec1d745e6eaf631b2c3303badfd8"} Feb 25 13:22:24 crc kubenswrapper[4815]: I0225 13:22:24.935442 4815 scope.go:117] "RemoveContainer" containerID="0bb59977532b467793fb68d57745c41d814b3da95fcaef8a163ec0c6cc34ce02" Feb 25 13:22:25 crc kubenswrapper[4815]: I0225 13:22:25.476963 4815 generic.go:334] "Generic (PLEG): container finished" podID="2c1f35ba-e05e-44ff-bdd4-bf725376b380" containerID="a97eed3aa89aada09e3e8712189831410ccc00e7beed4af2e5d11b73672f598f" exitCode=0 Feb 25 13:22:25 crc kubenswrapper[4815]: I0225 13:22:25.477048 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-k8vxf" event={"ID":"2c1f35ba-e05e-44ff-bdd4-bf725376b380","Type":"ContainerDied","Data":"a97eed3aa89aada09e3e8712189831410ccc00e7beed4af2e5d11b73672f598f"} Feb 25 13:22:25 crc kubenswrapper[4815]: I0225 13:22:25.480080 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 25 13:22:25 crc kubenswrapper[4815]: I0225 13:22:25.482787 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"75f8f6e705409d40c4115c32c4e36f96c90c0da9d5c56b92f4a0c524c12abe91"} Feb 25 13:22:25 crc kubenswrapper[4815]: I0225 13:22:25.483472 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:22:25 crc kubenswrapper[4815]: I0225 13:22:25.490027 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerStarted","Data":"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837"} Feb 25 13:22:25 crc kubenswrapper[4815]: I0225 13:22:25.490345 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:25 crc kubenswrapper[4815]: I0225 13:22:25.490362 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:25 crc kubenswrapper[4815]: I0225 13:22:25.490371 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:25 crc kubenswrapper[4815]: I0225 13:22:25.517398 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:25 crc kubenswrapper[4815]: I0225 13:22:25.518531 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:25 crc kubenswrapper[4815]: I0225 13:22:25.568631 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" podStartSLOduration=50.568614843 podStartE2EDuration="50.568614843s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:25.56787892 +0000 UTC m=+103.368976984" watchObservedRunningTime="2026-02-25 13:22:25.568614843 +0000 UTC m=+103.369712897" Feb 25 13:22:25 crc kubenswrapper[4815]: I0225 13:22:25.597455 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.597436597 podStartE2EDuration="16.597436597s" podCreationTimestamp="2026-02-25 13:22:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:25.597299824 +0000 UTC m=+103.398397878" watchObservedRunningTime="2026-02-25 13:22:25.597436597 +0000 UTC m=+103.398534641" Feb 25 13:22:25 crc kubenswrapper[4815]: I0225 13:22:25.935010 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:25 crc kubenswrapper[4815]: I0225 13:22:25.935010 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:25 crc kubenswrapper[4815]: E0225 13:22:25.935143 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 13:22:25 crc kubenswrapper[4815]: I0225 13:22:25.935058 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:25 crc kubenswrapper[4815]: E0225 13:22:25.935241 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 13:22:25 crc kubenswrapper[4815]: I0225 13:22:25.935272 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:25 crc kubenswrapper[4815]: E0225 13:22:25.935385 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbqbs" podUID="97b044d0-af42-4f42-95a8-49d6c8782246" Feb 25 13:22:25 crc kubenswrapper[4815]: E0225 13:22:25.935449 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 13:22:26 crc kubenswrapper[4815]: I0225 13:22:26.499743 4815 generic.go:334] "Generic (PLEG): container finished" podID="2c1f35ba-e05e-44ff-bdd4-bf725376b380" containerID="9534df36d13cbd77683e4161ffb4d3dfdb8da5be4d97067100eeed6ec0cae664" exitCode=0 Feb 25 13:22:26 crc kubenswrapper[4815]: I0225 13:22:26.499800 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-k8vxf" event={"ID":"2c1f35ba-e05e-44ff-bdd4-bf725376b380","Type":"ContainerDied","Data":"9534df36d13cbd77683e4161ffb4d3dfdb8da5be4d97067100eeed6ec0cae664"} Feb 25 13:22:27 crc kubenswrapper[4815]: I0225 13:22:27.510095 4815 generic.go:334] "Generic (PLEG): container finished" podID="2c1f35ba-e05e-44ff-bdd4-bf725376b380" containerID="4c2548f82100b49b4ce41b7218e31603f02c233d23f2405782da0f43cf716d47" exitCode=0 Feb 25 13:22:27 crc kubenswrapper[4815]: I0225 13:22:27.510652 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-k8vxf" event={"ID":"2c1f35ba-e05e-44ff-bdd4-bf725376b380","Type":"ContainerDied","Data":"4c2548f82100b49b4ce41b7218e31603f02c233d23f2405782da0f43cf716d47"} Feb 25 13:22:27 crc kubenswrapper[4815]: I0225 13:22:27.616054 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gbqbs"] Feb 25 13:22:27 crc kubenswrapper[4815]: I0225 13:22:27.616170 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:27 crc kubenswrapper[4815]: E0225 13:22:27.616277 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbqbs" podUID="97b044d0-af42-4f42-95a8-49d6c8782246" Feb 25 13:22:27 crc kubenswrapper[4815]: I0225 13:22:27.653840 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs\") pod \"network-metrics-daemon-gbqbs\" (UID: \"97b044d0-af42-4f42-95a8-49d6c8782246\") " pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:27 crc kubenswrapper[4815]: E0225 13:22:27.653995 4815 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 13:22:27 crc kubenswrapper[4815]: E0225 13:22:27.654660 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs podName:97b044d0-af42-4f42-95a8-49d6c8782246 nodeName:}" failed. No retries permitted until 2026-02-25 13:22:35.654637847 +0000 UTC m=+113.455735901 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs") pod "network-metrics-daemon-gbqbs" (UID: "97b044d0-af42-4f42-95a8-49d6c8782246") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 13:22:27 crc kubenswrapper[4815]: I0225 13:22:27.934639 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:27 crc kubenswrapper[4815]: I0225 13:22:27.934683 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:27 crc kubenswrapper[4815]: I0225 13:22:27.934753 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:27 crc kubenswrapper[4815]: E0225 13:22:27.934861 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 13:22:27 crc kubenswrapper[4815]: E0225 13:22:27.934968 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 13:22:27 crc kubenswrapper[4815]: E0225 13:22:27.935056 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 13:22:28 crc kubenswrapper[4815]: I0225 13:22:28.519705 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-k8vxf" event={"ID":"2c1f35ba-e05e-44ff-bdd4-bf725376b380","Type":"ContainerStarted","Data":"7acbf4b4e060a594be784b8d5a908cc6aac0c0607d1c1338e4011473cb861883"} Feb 25 13:22:28 crc kubenswrapper[4815]: I0225 13:22:28.550122 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-k8vxf" podStartSLOduration=53.550102408 podStartE2EDuration="53.550102408s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:28.549708676 +0000 UTC m=+106.350806770" watchObservedRunningTime="2026-02-25 13:22:28.550102408 +0000 UTC m=+106.351200472" Feb 25 13:22:28 crc kubenswrapper[4815]: I0225 13:22:28.935453 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:28 crc kubenswrapper[4815]: E0225 13:22:28.935604 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gbqbs" podUID="97b044d0-af42-4f42-95a8-49d6c8782246" Feb 25 13:22:29 crc kubenswrapper[4815]: I0225 13:22:29.934609 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:29 crc kubenswrapper[4815]: I0225 13:22:29.934683 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:29 crc kubenswrapper[4815]: I0225 13:22:29.934872 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:29 crc kubenswrapper[4815]: E0225 13:22:29.935220 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 13:22:29 crc kubenswrapper[4815]: E0225 13:22:29.935053 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 13:22:29 crc kubenswrapper[4815]: E0225 13:22:29.935272 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.659321 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.659767 4815 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.741179 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-8jnnl"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.742303 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8jnnl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.743326 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.743723 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.746890 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.746955 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.747181 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.747266 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.747294 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.747421 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5wvt"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.747822 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5wvt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.748616 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.748786 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.749029 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.749658 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.750055 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.751539 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.751713 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qscdn"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.752092 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qscdn" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.753146 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4klbc"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.753739 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.754491 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.754992 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.758715 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.763851 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vlx6g"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.764128 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6glz5"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.764491 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-5chwn"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.764749 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hckbl"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.765306 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.765814 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.766061 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.766932 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.768299 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.768625 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.770632 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fhfxk"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.771409 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fhfxk" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.771623 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.788447 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lpx8"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.788773 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.789085 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qkrlq"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.789918 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lpx8" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.790085 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qkrlq" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.799931 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.800001 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.800125 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.806353 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.806542 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.806614 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.806679 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.806740 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.806797 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.806910 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.807176 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.807913 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.811085 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-z4lg2"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.811539 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ff75l"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.811851 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ff75l" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.812139 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.812188 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.812552 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.812647 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.812698 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.812793 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.812870 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.812913 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.813020 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.813167 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.813932 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.814096 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.814245 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.814345 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.814478 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.818712 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.818739 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.818775 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.819009 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.819145 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-tswct"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.819330 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.819149 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.819484 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.819204 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.819236 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.819238 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.819363 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.820993 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dqj92"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.821638 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-dqj92" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.822624 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.822731 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.825219 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.827033 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.827116 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.827261 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.827318 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.827394 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.827455 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.827476 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.827703 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.828721 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.828839 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.828918 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.828717 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.829700 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.830828 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.831130 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.831307 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.831462 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.831659 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.831848 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.832024 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.832235 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.832434 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.833583 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.834110 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.834262 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.834291 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.834335 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.834432 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.834471 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.834491 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.834562 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.834611 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.834699 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.834822 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.835673 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.835801 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jcf7k"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.835880 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.836199 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.836287 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jcf7k" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.837792 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c699m"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.838532 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c699m" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.867500 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.868353 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.869160 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-mhgxd"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.869893 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.870248 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.873286 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.873823 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.874445 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.874930 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.875086 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.875312 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.878287 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.878469 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.923918 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.924201 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.924321 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.924793 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.924911 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.924973 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.925028 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.925222 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.925700 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.928456 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-md7bn"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.928983 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.929388 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.929720 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.929735 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.929940 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.930133 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.930696 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.931980 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e9ae59fe-6c8c-4906-8c36-2b808d0474c8-default-certificate\") pod \"router-default-5444994796-5chwn\" (UID: \"e9ae59fe-6c8c-4906-8c36-2b808d0474c8\") " pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932011 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cd91d14c-0e3a-48a8-948f-22248711def6-trusted-ca\") pod \"ingress-operator-5b745b69d9-qkx6p\" (UID: \"cd91d14c-0e3a-48a8-948f-22248711def6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932039 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2a76074e-1587-4795-a538-73d1fee5b028-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-6pxmc\" (UID: \"2a76074e-1587-4795-a538-73d1fee5b028\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932064 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mzqp\" (UniqueName: \"kubernetes.io/projected/556bab93-68c3-4771-87cf-107ba90525ea-kube-api-access-8mzqp\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932086 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6f28beca-7506-4390-9c74-86466a6b09ef-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6glz5\" (UID: \"6f28beca-7506-4390-9c74-86466a6b09ef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932112 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s59t\" (UniqueName: \"kubernetes.io/projected/6f28beca-7506-4390-9c74-86466a6b09ef-kube-api-access-4s59t\") pod \"machine-api-operator-5694c8668f-6glz5\" (UID: \"6f28beca-7506-4390-9c74-86466a6b09ef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932136 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2a76074e-1587-4795-a538-73d1fee5b028-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-6pxmc\" (UID: \"2a76074e-1587-4795-a538-73d1fee5b028\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932162 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mffbs\" (UniqueName: \"kubernetes.io/projected/e9ae59fe-6c8c-4906-8c36-2b808d0474c8-kube-api-access-mffbs\") pod \"router-default-5444994796-5chwn\" (UID: \"e9ae59fe-6c8c-4906-8c36-2b808d0474c8\") " pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932186 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5358560f-45f9-4548-8521-fad61a51f5a9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-z7rhl\" (UID: \"5358560f-45f9-4548-8521-fad61a51f5a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932212 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0e8fcc05-9af2-497a-8639-109eaa9e75ad-auth-proxy-config\") pod \"machine-config-operator-74547568cd-gzlvz\" (UID: \"0e8fcc05-9af2-497a-8639-109eaa9e75ad\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932233 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-etcd-serving-ca\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932253 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6f28beca-7506-4390-9c74-86466a6b09ef-images\") pod \"machine-api-operator-5694c8668f-6glz5\" (UID: \"6f28beca-7506-4390-9c74-86466a6b09ef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932273 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4828s\" (UniqueName: \"kubernetes.io/projected/a2f6b143-248f-4851-966e-93009db4876b-kube-api-access-4828s\") pod \"kube-storage-version-migrator-operator-b67b599dd-qscdn\" (UID: \"a2f6b143-248f-4851-966e-93009db4876b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qscdn" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932295 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf04f84d-250b-46fc-beea-9edd1c0c236a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qkrlq\" (UID: \"bf04f84d-250b-46fc-beea-9edd1c0c236a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qkrlq" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932315 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-encryption-config\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932338 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0281a35e-29e2-420d-97c4-dc5fb607569a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7lpx8\" (UID: \"0281a35e-29e2-420d-97c4-dc5fb607569a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lpx8" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932360 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-config\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932380 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cd91d14c-0e3a-48a8-948f-22248711def6-metrics-tls\") pod \"ingress-operator-5b745b69d9-qkx6p\" (UID: \"cd91d14c-0e3a-48a8-948f-22248711def6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932404 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-client-ca\") pod \"controller-manager-879f6c89f-4klbc\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932438 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0281a35e-29e2-420d-97c4-dc5fb607569a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7lpx8\" (UID: \"0281a35e-29e2-420d-97c4-dc5fb607569a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lpx8" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932461 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/556bab93-68c3-4771-87cf-107ba90525ea-config\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932482 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-etcd-client\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932522 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-config\") pod \"controller-manager-879f6c89f-4klbc\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932547 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f28beca-7506-4390-9c74-86466a6b09ef-config\") pod \"machine-api-operator-5694c8668f-6glz5\" (UID: \"6f28beca-7506-4390-9c74-86466a6b09ef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932569 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2f6b143-248f-4851-966e-93009db4876b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-qscdn\" (UID: \"a2f6b143-248f-4851-966e-93009db4876b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qscdn" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932589 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0e8fcc05-9af2-497a-8639-109eaa9e75ad-images\") pod \"machine-config-operator-74547568cd-gzlvz\" (UID: \"0e8fcc05-9af2-497a-8639-109eaa9e75ad\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932611 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e9ae59fe-6c8c-4906-8c36-2b808d0474c8-stats-auth\") pod \"router-default-5444994796-5chwn\" (UID: \"e9ae59fe-6c8c-4906-8c36-2b808d0474c8\") " pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932632 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2a76074e-1587-4795-a538-73d1fee5b028-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-6pxmc\" (UID: \"2a76074e-1587-4795-a538-73d1fee5b028\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932656 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf04f84d-250b-46fc-beea-9edd1c0c236a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qkrlq\" (UID: \"bf04f84d-250b-46fc-beea-9edd1c0c236a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qkrlq" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932698 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8f2s\" (UniqueName: \"kubernetes.io/projected/5f80bb4b-ee87-4454-9335-62e45c44c5f4-kube-api-access-f8f2s\") pod \"machine-approver-56656f9798-r4kmx\" (UID: \"5f80bb4b-ee87-4454-9335-62e45c44c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932727 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5f80bb4b-ee87-4454-9335-62e45c44c5f4-machine-approver-tls\") pod \"machine-approver-56656f9798-r4kmx\" (UID: \"5f80bb4b-ee87-4454-9335-62e45c44c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932747 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/556bab93-68c3-4771-87cf-107ba90525ea-etcd-client\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932768 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/61c629d8-0773-44ce-8919-b32d32aa5d11-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-x5wvt\" (UID: \"61c629d8-0773-44ce-8919-b32d32aa5d11\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5wvt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932794 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a62c4db-cacf-45bb-9a8d-caefb482153b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-ff75l\" (UID: \"6a62c4db-cacf-45bb-9a8d-caefb482153b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ff75l" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932817 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwc6p\" (UniqueName: \"kubernetes.io/projected/5358560f-45f9-4548-8521-fad61a51f5a9-kube-api-access-vwc6p\") pod \"openshift-config-operator-7777fb866f-z7rhl\" (UID: \"5358560f-45f9-4548-8521-fad61a51f5a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932838 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2f6b143-248f-4851-966e-93009db4876b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-qscdn\" (UID: \"a2f6b143-248f-4851-966e-93009db4876b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qscdn" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932859 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9ae59fe-6c8c-4906-8c36-2b808d0474c8-service-ca-bundle\") pod \"router-default-5444994796-5chwn\" (UID: \"e9ae59fe-6c8c-4906-8c36-2b808d0474c8\") " pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932880 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppmzr\" (UniqueName: \"kubernetes.io/projected/cd91d14c-0e3a-48a8-948f-22248711def6-kube-api-access-ppmzr\") pod \"ingress-operator-5b745b69d9-qkx6p\" (UID: \"cd91d14c-0e3a-48a8-948f-22248711def6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932900 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z49gr\" (UniqueName: \"kubernetes.io/projected/fb4ee7c5-3dbf-4256-83e8-482b44dfea31-kube-api-access-z49gr\") pod \"migrator-59844c95c7-fhfxk\" (UID: \"fb4ee7c5-3dbf-4256-83e8-482b44dfea31\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fhfxk" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932921 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d6t6\" (UniqueName: \"kubernetes.io/projected/60971790-c5dc-4884-ae44-e91d035cbe31-kube-api-access-2d6t6\") pod \"controller-manager-879f6c89f-4klbc\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932942 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f80bb4b-ee87-4454-9335-62e45c44c5f4-config\") pod \"machine-approver-56656f9798-r4kmx\" (UID: \"5f80bb4b-ee87-4454-9335-62e45c44c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932966 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0281a35e-29e2-420d-97c4-dc5fb607569a-config\") pod \"kube-controller-manager-operator-78b949d7b-7lpx8\" (UID: \"0281a35e-29e2-420d-97c4-dc5fb607569a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lpx8" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.932987 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsp8j\" (UniqueName: \"kubernetes.io/projected/61c629d8-0773-44ce-8919-b32d32aa5d11-kube-api-access-nsp8j\") pod \"control-plane-machine-set-operator-78cbb6b69f-x5wvt\" (UID: \"61c629d8-0773-44ce-8919-b32d32aa5d11\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5wvt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933011 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkdkn\" (UniqueName: \"kubernetes.io/projected/bf04f84d-250b-46fc-beea-9edd1c0c236a-kube-api-access-hkdkn\") pod \"openshift-apiserver-operator-796bbdcf4f-qkrlq\" (UID: \"bf04f84d-250b-46fc-beea-9edd1c0c236a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qkrlq" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933031 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-audit-dir\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933050 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60971790-c5dc-4884-ae44-e91d035cbe31-serving-cert\") pod \"controller-manager-879f6c89f-4klbc\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933069 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5f80bb4b-ee87-4454-9335-62e45c44c5f4-auth-proxy-config\") pod \"machine-approver-56656f9798-r4kmx\" (UID: \"5f80bb4b-ee87-4454-9335-62e45c44c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933090 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-image-import-ca\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933111 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/556bab93-68c3-4771-87cf-107ba90525ea-serving-cert\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933132 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-node-pullsecrets\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933153 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cd91d14c-0e3a-48a8-948f-22248711def6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qkx6p\" (UID: \"cd91d14c-0e3a-48a8-948f-22248711def6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933185 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9ae59fe-6c8c-4906-8c36-2b808d0474c8-metrics-certs\") pod \"router-default-5444994796-5chwn\" (UID: \"e9ae59fe-6c8c-4906-8c36-2b808d0474c8\") " pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933208 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/556bab93-68c3-4771-87cf-107ba90525ea-etcd-ca\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933229 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-audit\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933247 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-serving-cert\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933269 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfgsx\" (UniqueName: \"kubernetes.io/projected/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-kube-api-access-sfgsx\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933288 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a62c4db-cacf-45bb-9a8d-caefb482153b-config\") pod \"kube-apiserver-operator-766d6c64bb-ff75l\" (UID: \"6a62c4db-cacf-45bb-9a8d-caefb482153b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ff75l" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933308 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6a62c4db-cacf-45bb-9a8d-caefb482153b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-ff75l\" (UID: \"6a62c4db-cacf-45bb-9a8d-caefb482153b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ff75l" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933329 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp5w9\" (UniqueName: \"kubernetes.io/projected/78d30b59-0a66-44e5-b0b9-a2864e4d4552-kube-api-access-lp5w9\") pod \"downloads-7954f5f757-8jnnl\" (UID: \"78d30b59-0a66-44e5-b0b9-a2864e4d4552\") " pod="openshift-console/downloads-7954f5f757-8jnnl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933353 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzf4j\" (UniqueName: \"kubernetes.io/projected/0e8fcc05-9af2-497a-8639-109eaa9e75ad-kube-api-access-kzf4j\") pod \"machine-config-operator-74547568cd-gzlvz\" (UID: \"0e8fcc05-9af2-497a-8639-109eaa9e75ad\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933372 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/556bab93-68c3-4771-87cf-107ba90525ea-etcd-service-ca\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933411 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0e8fcc05-9af2-497a-8639-109eaa9e75ad-proxy-tls\") pod \"machine-config-operator-74547568cd-gzlvz\" (UID: \"0e8fcc05-9af2-497a-8639-109eaa9e75ad\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933432 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d5sq\" (UniqueName: \"kubernetes.io/projected/2a76074e-1587-4795-a538-73d1fee5b028-kube-api-access-2d5sq\") pod \"cluster-image-registry-operator-dc59b4c8b-6pxmc\" (UID: \"2a76074e-1587-4795-a538-73d1fee5b028\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933453 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5358560f-45f9-4548-8521-fad61a51f5a9-serving-cert\") pod \"openshift-config-operator-7777fb866f-z7rhl\" (UID: \"5358560f-45f9-4548-8521-fad61a51f5a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933473 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.933495 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4klbc\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.940126 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.941115 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.941985 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.943536 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.943739 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.944409 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.944816 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.949970 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-njtsc"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.950809 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-njtsc" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.951005 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6th6z"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.951346 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.952939 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wdwxg"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.953605 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533762-lkzv9"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.953895 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.954269 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.954520 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wdwxg" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.954620 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533762-lkzv9" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.955017 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-w8lg6"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.955547 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-shdc4"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.956017 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-shdc4" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.956177 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-w8lg6" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.957077 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7g5t4"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.957561 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7g5t4" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.957817 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9t9rg"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.958578 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-9t9rg" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.960909 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8jnnl"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.961759 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qscdn"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.962916 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4klbc"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.964235 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-dv7nv"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.964928 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.965471 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5wvt"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.965570 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.966579 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.968845 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.970003 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6glz5"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.971550 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fhfxk"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.974858 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vlx6g"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.976604 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lpx8"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.979258 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qkrlq"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.982426 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.983328 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.984438 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.985005 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-tswct"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.986425 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.990396 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hckbl"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.993737 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.994305 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.995583 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6th6z"] Feb 25 13:22:30 crc kubenswrapper[4815]: I0225 13:22:30.997717 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mhgxd"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.000288 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c699m"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.001247 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ff75l"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.002167 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dqj92"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.003107 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-kw6xm"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.003880 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-kw6xm" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.005255 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.006207 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.007241 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wdwxg"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.008212 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jcf7k"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.009188 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.010268 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.011248 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533762-lkzv9"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.012183 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.013183 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-w8lg6"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.014155 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-njtsc"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.015082 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-md7bn"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.016075 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-z4lg2"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.016982 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-ccjcb"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.017404 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ccjcb" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.017940 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9t9rg"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.018916 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-kw6xm"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.019908 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-shdc4"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.020888 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7g5t4"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.021759 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-2dz2z"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.022220 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2dz2z" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.022687 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2dz2z"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.025557 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.031193 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-j9qns"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.032832 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.034777 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/556bab93-68c3-4771-87cf-107ba90525ea-etcd-service-ca\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.034816 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0e8fcc05-9af2-497a-8639-109eaa9e75ad-proxy-tls\") pod \"machine-config-operator-74547568cd-gzlvz\" (UID: \"0e8fcc05-9af2-497a-8639-109eaa9e75ad\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.034835 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d5sq\" (UniqueName: \"kubernetes.io/projected/2a76074e-1587-4795-a538-73d1fee5b028-kube-api-access-2d5sq\") pod \"cluster-image-registry-operator-dc59b4c8b-6pxmc\" (UID: \"2a76074e-1587-4795-a538-73d1fee5b028\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.034857 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.034881 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5358560f-45f9-4548-8521-fad61a51f5a9-serving-cert\") pod \"openshift-config-operator-7777fb866f-z7rhl\" (UID: \"5358560f-45f9-4548-8521-fad61a51f5a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.034899 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4klbc\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.034918 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e9ae59fe-6c8c-4906-8c36-2b808d0474c8-default-certificate\") pod \"router-default-5444994796-5chwn\" (UID: \"e9ae59fe-6c8c-4906-8c36-2b808d0474c8\") " pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.034933 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cd91d14c-0e3a-48a8-948f-22248711def6-trusted-ca\") pod \"ingress-operator-5b745b69d9-qkx6p\" (UID: \"cd91d14c-0e3a-48a8-948f-22248711def6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.034950 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2a76074e-1587-4795-a538-73d1fee5b028-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-6pxmc\" (UID: \"2a76074e-1587-4795-a538-73d1fee5b028\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.034966 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mzqp\" (UniqueName: \"kubernetes.io/projected/556bab93-68c3-4771-87cf-107ba90525ea-kube-api-access-8mzqp\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.034983 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6f28beca-7506-4390-9c74-86466a6b09ef-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6glz5\" (UID: \"6f28beca-7506-4390-9c74-86466a6b09ef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035000 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s59t\" (UniqueName: \"kubernetes.io/projected/6f28beca-7506-4390-9c74-86466a6b09ef-kube-api-access-4s59t\") pod \"machine-api-operator-5694c8668f-6glz5\" (UID: \"6f28beca-7506-4390-9c74-86466a6b09ef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035016 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2a76074e-1587-4795-a538-73d1fee5b028-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-6pxmc\" (UID: \"2a76074e-1587-4795-a538-73d1fee5b028\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035034 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mffbs\" (UniqueName: \"kubernetes.io/projected/e9ae59fe-6c8c-4906-8c36-2b808d0474c8-kube-api-access-mffbs\") pod \"router-default-5444994796-5chwn\" (UID: \"e9ae59fe-6c8c-4906-8c36-2b808d0474c8\") " pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035053 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5358560f-45f9-4548-8521-fad61a51f5a9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-z7rhl\" (UID: \"5358560f-45f9-4548-8521-fad61a51f5a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035073 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0e8fcc05-9af2-497a-8639-109eaa9e75ad-auth-proxy-config\") pod \"machine-config-operator-74547568cd-gzlvz\" (UID: \"0e8fcc05-9af2-497a-8639-109eaa9e75ad\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035088 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-etcd-serving-ca\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035104 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6f28beca-7506-4390-9c74-86466a6b09ef-images\") pod \"machine-api-operator-5694c8668f-6glz5\" (UID: \"6f28beca-7506-4390-9c74-86466a6b09ef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035119 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4828s\" (UniqueName: \"kubernetes.io/projected/a2f6b143-248f-4851-966e-93009db4876b-kube-api-access-4828s\") pod \"kube-storage-version-migrator-operator-b67b599dd-qscdn\" (UID: \"a2f6b143-248f-4851-966e-93009db4876b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qscdn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035136 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf04f84d-250b-46fc-beea-9edd1c0c236a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qkrlq\" (UID: \"bf04f84d-250b-46fc-beea-9edd1c0c236a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qkrlq" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035151 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-encryption-config\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035169 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0281a35e-29e2-420d-97c4-dc5fb607569a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7lpx8\" (UID: \"0281a35e-29e2-420d-97c4-dc5fb607569a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lpx8" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035186 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-config\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035201 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cd91d14c-0e3a-48a8-948f-22248711def6-metrics-tls\") pod \"ingress-operator-5b745b69d9-qkx6p\" (UID: \"cd91d14c-0e3a-48a8-948f-22248711def6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035216 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-client-ca\") pod \"controller-manager-879f6c89f-4klbc\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035240 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0281a35e-29e2-420d-97c4-dc5fb607569a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7lpx8\" (UID: \"0281a35e-29e2-420d-97c4-dc5fb607569a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lpx8" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035258 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/556bab93-68c3-4771-87cf-107ba90525ea-config\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035272 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-etcd-client\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035287 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-config\") pod \"controller-manager-879f6c89f-4klbc\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035303 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f28beca-7506-4390-9c74-86466a6b09ef-config\") pod \"machine-api-operator-5694c8668f-6glz5\" (UID: \"6f28beca-7506-4390-9c74-86466a6b09ef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035319 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2f6b143-248f-4851-966e-93009db4876b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-qscdn\" (UID: \"a2f6b143-248f-4851-966e-93009db4876b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qscdn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035335 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0e8fcc05-9af2-497a-8639-109eaa9e75ad-images\") pod \"machine-config-operator-74547568cd-gzlvz\" (UID: \"0e8fcc05-9af2-497a-8639-109eaa9e75ad\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035349 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e9ae59fe-6c8c-4906-8c36-2b808d0474c8-stats-auth\") pod \"router-default-5444994796-5chwn\" (UID: \"e9ae59fe-6c8c-4906-8c36-2b808d0474c8\") " pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035365 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2a76074e-1587-4795-a538-73d1fee5b028-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-6pxmc\" (UID: \"2a76074e-1587-4795-a538-73d1fee5b028\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035386 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf04f84d-250b-46fc-beea-9edd1c0c236a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qkrlq\" (UID: \"bf04f84d-250b-46fc-beea-9edd1c0c236a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qkrlq" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035465 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8f2s\" (UniqueName: \"kubernetes.io/projected/5f80bb4b-ee87-4454-9335-62e45c44c5f4-kube-api-access-f8f2s\") pod \"machine-approver-56656f9798-r4kmx\" (UID: \"5f80bb4b-ee87-4454-9335-62e45c44c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035483 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/556bab93-68c3-4771-87cf-107ba90525ea-etcd-client\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.035800 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5f80bb4b-ee87-4454-9335-62e45c44c5f4-machine-approver-tls\") pod \"machine-approver-56656f9798-r4kmx\" (UID: \"5f80bb4b-ee87-4454-9335-62e45c44c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.036822 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5358560f-45f9-4548-8521-fad61a51f5a9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-z7rhl\" (UID: \"5358560f-45f9-4548-8521-fad61a51f5a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.036978 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-config\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.036988 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/61c629d8-0773-44ce-8919-b32d32aa5d11-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-x5wvt\" (UID: \"61c629d8-0773-44ce-8919-b32d32aa5d11\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5wvt" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037067 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a62c4db-cacf-45bb-9a8d-caefb482153b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-ff75l\" (UID: \"6a62c4db-cacf-45bb-9a8d-caefb482153b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ff75l" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037066 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2f6b143-248f-4851-966e-93009db4876b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-qscdn\" (UID: \"a2f6b143-248f-4851-966e-93009db4876b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qscdn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037087 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwc6p\" (UniqueName: \"kubernetes.io/projected/5358560f-45f9-4548-8521-fad61a51f5a9-kube-api-access-vwc6p\") pod \"openshift-config-operator-7777fb866f-z7rhl\" (UID: \"5358560f-45f9-4548-8521-fad61a51f5a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037108 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2f6b143-248f-4851-966e-93009db4876b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-qscdn\" (UID: \"a2f6b143-248f-4851-966e-93009db4876b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qscdn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037130 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0281a35e-29e2-420d-97c4-dc5fb607569a-config\") pod \"kube-controller-manager-operator-78b949d7b-7lpx8\" (UID: \"0281a35e-29e2-420d-97c4-dc5fb607569a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lpx8" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037148 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9ae59fe-6c8c-4906-8c36-2b808d0474c8-service-ca-bundle\") pod \"router-default-5444994796-5chwn\" (UID: \"e9ae59fe-6c8c-4906-8c36-2b808d0474c8\") " pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037166 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppmzr\" (UniqueName: \"kubernetes.io/projected/cd91d14c-0e3a-48a8-948f-22248711def6-kube-api-access-ppmzr\") pod \"ingress-operator-5b745b69d9-qkx6p\" (UID: \"cd91d14c-0e3a-48a8-948f-22248711def6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037186 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z49gr\" (UniqueName: \"kubernetes.io/projected/fb4ee7c5-3dbf-4256-83e8-482b44dfea31-kube-api-access-z49gr\") pod \"migrator-59844c95c7-fhfxk\" (UID: \"fb4ee7c5-3dbf-4256-83e8-482b44dfea31\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fhfxk" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037204 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d6t6\" (UniqueName: \"kubernetes.io/projected/60971790-c5dc-4884-ae44-e91d035cbe31-kube-api-access-2d6t6\") pod \"controller-manager-879f6c89f-4klbc\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037221 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f80bb4b-ee87-4454-9335-62e45c44c5f4-config\") pod \"machine-approver-56656f9798-r4kmx\" (UID: \"5f80bb4b-ee87-4454-9335-62e45c44c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037238 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsp8j\" (UniqueName: \"kubernetes.io/projected/61c629d8-0773-44ce-8919-b32d32aa5d11-kube-api-access-nsp8j\") pod \"control-plane-machine-set-operator-78cbb6b69f-x5wvt\" (UID: \"61c629d8-0773-44ce-8919-b32d32aa5d11\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5wvt" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037255 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkdkn\" (UniqueName: \"kubernetes.io/projected/bf04f84d-250b-46fc-beea-9edd1c0c236a-kube-api-access-hkdkn\") pod \"openshift-apiserver-operator-796bbdcf4f-qkrlq\" (UID: \"bf04f84d-250b-46fc-beea-9edd1c0c236a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qkrlq" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037278 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-audit-dir\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037294 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60971790-c5dc-4884-ae44-e91d035cbe31-serving-cert\") pod \"controller-manager-879f6c89f-4klbc\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037310 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5f80bb4b-ee87-4454-9335-62e45c44c5f4-auth-proxy-config\") pod \"machine-approver-56656f9798-r4kmx\" (UID: \"5f80bb4b-ee87-4454-9335-62e45c44c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037328 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-image-import-ca\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037362 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/556bab93-68c3-4771-87cf-107ba90525ea-serving-cert\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037385 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-node-pullsecrets\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037408 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cd91d14c-0e3a-48a8-948f-22248711def6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qkx6p\" (UID: \"cd91d14c-0e3a-48a8-948f-22248711def6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037449 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9ae59fe-6c8c-4906-8c36-2b808d0474c8-metrics-certs\") pod \"router-default-5444994796-5chwn\" (UID: \"e9ae59fe-6c8c-4906-8c36-2b808d0474c8\") " pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037482 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/556bab93-68c3-4771-87cf-107ba90525ea-etcd-ca\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037528 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzf4j\" (UniqueName: \"kubernetes.io/projected/0e8fcc05-9af2-497a-8639-109eaa9e75ad-kube-api-access-kzf4j\") pod \"machine-config-operator-74547568cd-gzlvz\" (UID: \"0e8fcc05-9af2-497a-8639-109eaa9e75ad\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037547 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-audit\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037563 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-serving-cert\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037579 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfgsx\" (UniqueName: \"kubernetes.io/projected/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-kube-api-access-sfgsx\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037598 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a62c4db-cacf-45bb-9a8d-caefb482153b-config\") pod \"kube-apiserver-operator-766d6c64bb-ff75l\" (UID: \"6a62c4db-cacf-45bb-9a8d-caefb482153b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ff75l" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037615 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6a62c4db-cacf-45bb-9a8d-caefb482153b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-ff75l\" (UID: \"6a62c4db-cacf-45bb-9a8d-caefb482153b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ff75l" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037634 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp5w9\" (UniqueName: \"kubernetes.io/projected/78d30b59-0a66-44e5-b0b9-a2864e4d4552-kube-api-access-lp5w9\") pod \"downloads-7954f5f757-8jnnl\" (UID: \"78d30b59-0a66-44e5-b0b9-a2864e4d4552\") " pod="openshift-console/downloads-7954f5f757-8jnnl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037640 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0e8fcc05-9af2-497a-8639-109eaa9e75ad-auth-proxy-config\") pod \"machine-config-operator-74547568cd-gzlvz\" (UID: \"0e8fcc05-9af2-497a-8639-109eaa9e75ad\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037681 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4klbc\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037846 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-j9qns"] Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.038010 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-etcd-serving-ca\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.038237 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0e8fcc05-9af2-497a-8639-109eaa9e75ad-images\") pod \"machine-config-operator-74547568cd-gzlvz\" (UID: \"0e8fcc05-9af2-497a-8639-109eaa9e75ad\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.038260 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cd91d14c-0e3a-48a8-948f-22248711def6-trusted-ca\") pod \"ingress-operator-5b745b69d9-qkx6p\" (UID: \"cd91d14c-0e3a-48a8-948f-22248711def6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.038310 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/556bab93-68c3-4771-87cf-107ba90525ea-etcd-service-ca\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.038681 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6f28beca-7506-4390-9c74-86466a6b09ef-images\") pod \"machine-api-operator-5694c8668f-6glz5\" (UID: \"6f28beca-7506-4390-9c74-86466a6b09ef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.039082 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-audit-dir\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.039130 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2a76074e-1587-4795-a538-73d1fee5b028-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-6pxmc\" (UID: \"2a76074e-1587-4795-a538-73d1fee5b028\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.039135 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/556bab93-68c3-4771-87cf-107ba90525ea-config\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.037449 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.040668 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f28beca-7506-4390-9c74-86466a6b09ef-config\") pod \"machine-api-operator-5694c8668f-6glz5\" (UID: \"6f28beca-7506-4390-9c74-86466a6b09ef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.040800 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5f80bb4b-ee87-4454-9335-62e45c44c5f4-auth-proxy-config\") pod \"machine-approver-56656f9798-r4kmx\" (UID: \"5f80bb4b-ee87-4454-9335-62e45c44c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.041229 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-config\") pod \"controller-manager-879f6c89f-4klbc\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.041545 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-client-ca\") pod \"controller-manager-879f6c89f-4klbc\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.041596 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf04f84d-250b-46fc-beea-9edd1c0c236a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qkrlq\" (UID: \"bf04f84d-250b-46fc-beea-9edd1c0c236a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qkrlq" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.041627 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-node-pullsecrets\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.041696 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-image-import-ca\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.042262 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a62c4db-cacf-45bb-9a8d-caefb482153b-config\") pod \"kube-apiserver-operator-766d6c64bb-ff75l\" (UID: \"6a62c4db-cacf-45bb-9a8d-caefb482153b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ff75l" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.042479 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f80bb4b-ee87-4454-9335-62e45c44c5f4-config\") pod \"machine-approver-56656f9798-r4kmx\" (UID: \"5f80bb4b-ee87-4454-9335-62e45c44c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.042673 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9ae59fe-6c8c-4906-8c36-2b808d0474c8-service-ca-bundle\") pod \"router-default-5444994796-5chwn\" (UID: \"e9ae59fe-6c8c-4906-8c36-2b808d0474c8\") " pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.042707 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0281a35e-29e2-420d-97c4-dc5fb607569a-config\") pod \"kube-controller-manager-operator-78b949d7b-7lpx8\" (UID: \"0281a35e-29e2-420d-97c4-dc5fb607569a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lpx8" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.042859 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/556bab93-68c3-4771-87cf-107ba90525ea-etcd-ca\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.042867 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-audit\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.044229 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.044257 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf04f84d-250b-46fc-beea-9edd1c0c236a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qkrlq\" (UID: \"bf04f84d-250b-46fc-beea-9edd1c0c236a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qkrlq" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.044288 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e9ae59fe-6c8c-4906-8c36-2b808d0474c8-default-certificate\") pod \"router-default-5444994796-5chwn\" (UID: \"e9ae59fe-6c8c-4906-8c36-2b808d0474c8\") " pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.044343 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5358560f-45f9-4548-8521-fad61a51f5a9-serving-cert\") pod \"openshift-config-operator-7777fb866f-z7rhl\" (UID: \"5358560f-45f9-4548-8521-fad61a51f5a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.044967 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e9ae59fe-6c8c-4906-8c36-2b808d0474c8-stats-auth\") pod \"router-default-5444994796-5chwn\" (UID: \"e9ae59fe-6c8c-4906-8c36-2b808d0474c8\") " pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.045270 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-serving-cert\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.045556 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/61c629d8-0773-44ce-8919-b32d32aa5d11-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-x5wvt\" (UID: \"61c629d8-0773-44ce-8919-b32d32aa5d11\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5wvt" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.046248 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-etcd-client\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.046703 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0281a35e-29e2-420d-97c4-dc5fb607569a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7lpx8\" (UID: \"0281a35e-29e2-420d-97c4-dc5fb607569a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lpx8" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.047397 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-encryption-config\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.047603 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cd91d14c-0e3a-48a8-948f-22248711def6-metrics-tls\") pod \"ingress-operator-5b745b69d9-qkx6p\" (UID: \"cd91d14c-0e3a-48a8-948f-22248711def6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.047850 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2f6b143-248f-4851-966e-93009db4876b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-qscdn\" (UID: \"a2f6b143-248f-4851-966e-93009db4876b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qscdn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.047918 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2a76074e-1587-4795-a538-73d1fee5b028-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-6pxmc\" (UID: \"2a76074e-1587-4795-a538-73d1fee5b028\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.047984 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6f28beca-7506-4390-9c74-86466a6b09ef-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6glz5\" (UID: \"6f28beca-7506-4390-9c74-86466a6b09ef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.048150 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0e8fcc05-9af2-497a-8639-109eaa9e75ad-proxy-tls\") pod \"machine-config-operator-74547568cd-gzlvz\" (UID: \"0e8fcc05-9af2-497a-8639-109eaa9e75ad\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.048418 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a62c4db-cacf-45bb-9a8d-caefb482153b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-ff75l\" (UID: \"6a62c4db-cacf-45bb-9a8d-caefb482153b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ff75l" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.048706 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60971790-c5dc-4884-ae44-e91d035cbe31-serving-cert\") pod \"controller-manager-879f6c89f-4klbc\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.048839 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5f80bb4b-ee87-4454-9335-62e45c44c5f4-machine-approver-tls\") pod \"machine-approver-56656f9798-r4kmx\" (UID: \"5f80bb4b-ee87-4454-9335-62e45c44c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.048951 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9ae59fe-6c8c-4906-8c36-2b808d0474c8-metrics-certs\") pod \"router-default-5444994796-5chwn\" (UID: \"e9ae59fe-6c8c-4906-8c36-2b808d0474c8\") " pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.049394 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/556bab93-68c3-4771-87cf-107ba90525ea-etcd-client\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.049427 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/556bab93-68c3-4771-87cf-107ba90525ea-serving-cert\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.064120 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.085191 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.105206 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.126661 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.144951 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.164944 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.185409 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.205783 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.225089 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.245220 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.265198 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.285656 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.304276 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.325069 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.345753 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.366296 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.393281 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.405041 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.425993 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.445242 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.465486 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.485876 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.506756 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.526332 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.565618 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.595994 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.606357 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.626378 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.643795 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-registry-tls\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.643949 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2l6w\" (UniqueName: \"kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-kube-api-access-x2l6w\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.644045 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-bound-sa-token\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.644253 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/072fbe81-88e1-4755-bb59-5ee95d72760f-trusted-ca\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.644382 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/072fbe81-88e1-4755-bb59-5ee95d72760f-registry-certificates\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.644454 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/072fbe81-88e1-4755-bb59-5ee95d72760f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.644670 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/072fbe81-88e1-4755-bb59-5ee95d72760f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.644744 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: E0225 13:22:31.645249 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:32.145223154 +0000 UTC m=+109.946321308 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.646041 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.673289 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.686302 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.705755 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.726276 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.745975 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.746189 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/739241b1-c772-4ef7-b149-e2098da43d67-webhook-cert\") pod \"packageserver-d55dfcdfc-h7bdq\" (UID: \"739241b1-c772-4ef7-b149-e2098da43d67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.746227 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3c08c2da-42cd-47ba-92e6-aed1d48855d3-csi-data-dir\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.746283 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7rzd\" (UniqueName: \"kubernetes.io/projected/71066f15-4379-475e-8295-671bfd6ea42b-kube-api-access-q7rzd\") pod \"catalog-operator-68c6474976-lttkb\" (UID: \"71066f15-4379-475e-8295-671bfd6ea42b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.746311 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-encryption-config\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.746342 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dv4r\" (UniqueName: \"kubernetes.io/projected/bdc0843c-6110-4d69-b444-7e5384f858e1-kube-api-access-7dv4r\") pod \"console-operator-58897d9998-w8lg6\" (UID: \"bdc0843c-6110-4d69-b444-7e5384f858e1\") " pod="openshift-console-operator/console-operator-58897d9998-w8lg6" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.746371 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86tlk\" (UniqueName: \"kubernetes.io/projected/887d2231-1877-4977-ac7d-9f7cf1b33a23-kube-api-access-86tlk\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.746393 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-etcd-client\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: E0225 13:22:31.746590 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:32.24649005 +0000 UTC m=+110.047588134 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.746798 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.746888 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-oauth-config\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.746967 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72d966ad-816d-4483-b9e6-44edb10f9d75-config\") pod \"route-controller-manager-6576b87f9c-plsdd\" (UID: \"72d966ad-816d-4483-b9e6-44edb10f9d75\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.747074 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.747458 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlj7n\" (UniqueName: \"kubernetes.io/projected/ebefeb1c-aa64-4a51-8d2d-197d8e8421af-kube-api-access-mlj7n\") pod \"openshift-controller-manager-operator-756b6f6bc6-jcf7k\" (UID: \"ebefeb1c-aa64-4a51-8d2d-197d8e8421af\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jcf7k" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.747626 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfh4w\" (UniqueName: \"kubernetes.io/projected/3c08c2da-42cd-47ba-92e6-aed1d48855d3-kube-api-access-lfh4w\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.747725 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-service-ca\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.747838 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/770461e1-2b0c-4e47-90d1-67f2286e88e2-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-w2rk4\" (UID: \"770461e1-2b0c-4e47-90d1-67f2286e88e2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.748054 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp9x9\" (UniqueName: \"kubernetes.io/projected/601c110b-f7ba-4f1c-92ca-dc3cc6634b53-kube-api-access-zp9x9\") pod \"authentication-operator-69f744f599-tswct\" (UID: \"601c110b-f7ba-4f1c-92ca-dc3cc6634b53\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.748153 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-audit-policies\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: E0225 13:22:31.748350 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:32.248208601 +0000 UTC m=+110.049306705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.748571 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.748913 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87c350ba-fd2f-4e8c-a057-be035865dbf4-serving-cert\") pod \"service-ca-operator-777779d784-7g5t4\" (UID: \"87c350ba-fd2f-4e8c-a057-be035865dbf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7g5t4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.749454 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs2dt\" (UniqueName: \"kubernetes.io/projected/3503ed80-0534-4d0b-a69a-fc9e9c5fa092-kube-api-access-fs2dt\") pod \"olm-operator-6b444d44fb-qdgb5\" (UID: \"3503ed80-0534-4d0b-a69a-fc9e9c5fa092\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.749561 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmrn9\" (UniqueName: \"kubernetes.io/projected/91bd28b3-4940-4027-8753-18335157566e-kube-api-access-vmrn9\") pod \"dns-operator-744455d44c-9t9rg\" (UID: \"91bd28b3-4940-4027-8753-18335157566e\") " pod="openshift-dns-operator/dns-operator-744455d44c-9t9rg" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.750022 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/643cac92-ba9e-4d3c-97b7-ca525f8fc046-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c699m\" (UID: \"643cac92-ba9e-4d3c-97b7-ca525f8fc046\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c699m" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.750108 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.750176 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k82t\" (UniqueName: \"kubernetes.io/projected/6403af09-14ef-49fc-858f-06c1bbadb88b-kube-api-access-7k82t\") pod \"auto-csr-approver-29533762-lkzv9\" (UID: \"6403af09-14ef-49fc-858f-06c1bbadb88b\") " pod="openshift-infra/auto-csr-approver-29533762-lkzv9" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.750257 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6zdr\" (UniqueName: \"kubernetes.io/projected/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-kube-api-access-c6zdr\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.750304 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggzb8\" (UniqueName: \"kubernetes.io/projected/65dfccf6-aa91-41fa-95d6-fd4778065408-kube-api-access-ggzb8\") pod \"cni-sysctl-allowlist-ds-dv7nv\" (UID: \"65dfccf6-aa91-41fa-95d6-fd4778065408\") " pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.750381 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr9jx\" (UniqueName: \"kubernetes.io/projected/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-kube-api-access-hr9jx\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.750426 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdc0843c-6110-4d69-b444-7e5384f858e1-config\") pod \"console-operator-58897d9998-w8lg6\" (UID: \"bdc0843c-6110-4d69-b444-7e5384f858e1\") " pod="openshift-console-operator/console-operator-58897d9998-w8lg6" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.750538 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d6db35c3-7baa-4452-b8ff-5c20470142d6-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dqj92\" (UID: \"d6db35c3-7baa-4452-b8ff-5c20470142d6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dqj92" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.750595 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-trusted-ca-bundle\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.750649 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brbmt\" (UniqueName: \"kubernetes.io/projected/72d966ad-816d-4483-b9e6-44edb10f9d75-kube-api-access-brbmt\") pod \"route-controller-manager-6576b87f9c-plsdd\" (UID: \"72d966ad-816d-4483-b9e6-44edb10f9d75\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.750696 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/739241b1-c772-4ef7-b149-e2098da43d67-tmpfs\") pod \"packageserver-d55dfcdfc-h7bdq\" (UID: \"739241b1-c772-4ef7-b149-e2098da43d67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.750797 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/08e7cd4c-1c51-4142-9384-c0d4561e7e36-signing-cabundle\") pod \"service-ca-9c57cc56f-wdwxg\" (UID: \"08e7cd4c-1c51-4142-9384-c0d4561e7e36\") " pod="openshift-service-ca/service-ca-9c57cc56f-wdwxg" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.750842 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bdc0843c-6110-4d69-b444-7e5384f858e1-trusted-ca\") pod \"console-operator-58897d9998-w8lg6\" (UID: \"bdc0843c-6110-4d69-b444-7e5384f858e1\") " pod="openshift-console-operator/console-operator-58897d9998-w8lg6" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.750892 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksc96\" (UniqueName: \"kubernetes.io/projected/7bcf8b34-940d-41d9-b32a-5548c5cc62c2-kube-api-access-ksc96\") pod \"package-server-manager-789f6589d5-njtsc\" (UID: \"7bcf8b34-940d-41d9-b32a-5548c5cc62c2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-njtsc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.750982 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km5v7\" (UniqueName: \"kubernetes.io/projected/1b2fb1a6-e1d1-48e0-83dc-17143971294e-kube-api-access-km5v7\") pod \"marketplace-operator-79b997595-6th6z\" (UID: \"1b2fb1a6-e1d1-48e0-83dc-17143971294e\") " pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.751347 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-audit-dir\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.751436 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.751550 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3c08c2da-42cd-47ba-92e6-aed1d48855d3-registration-dir\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.751614 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdvqb\" (UniqueName: \"kubernetes.io/projected/770461e1-2b0c-4e47-90d1-67f2286e88e2-kube-api-access-qdvqb\") pod \"machine-config-controller-84d6567774-w2rk4\" (UID: \"770461e1-2b0c-4e47-90d1-67f2286e88e2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.751663 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/65dfccf6-aa91-41fa-95d6-fd4778065408-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-dv7nv\" (UID: \"65dfccf6-aa91-41fa-95d6-fd4778065408\") " pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.751817 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebefeb1c-aa64-4a51-8d2d-197d8e8421af-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jcf7k\" (UID: \"ebefeb1c-aa64-4a51-8d2d-197d8e8421af\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jcf7k" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.751862 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-serving-cert\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.751904 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-config\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.751965 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84b7af84-2b2c-48f7-b7b0-e38fb0f7d702-config-volume\") pod \"dns-default-kw6xm\" (UID: \"84b7af84-2b2c-48f7-b7b0-e38fb0f7d702\") " pod="openshift-dns/dns-default-kw6xm" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752057 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/072fbe81-88e1-4755-bb59-5ee95d72760f-trusted-ca\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752095 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bcf8b34-940d-41d9-b32a-5548c5cc62c2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-njtsc\" (UID: \"7bcf8b34-940d-41d9-b32a-5548c5cc62c2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-njtsc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752130 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752194 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1b2fb1a6-e1d1-48e0-83dc-17143971294e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6th6z\" (UID: \"1b2fb1a6-e1d1-48e0-83dc-17143971294e\") " pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752224 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/601c110b-f7ba-4f1c-92ca-dc3cc6634b53-service-ca-bundle\") pod \"authentication-operator-69f744f599-tswct\" (UID: \"601c110b-f7ba-4f1c-92ca-dc3cc6634b53\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752253 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-config-volume\") pod \"collect-profiles-29533755-65vjv\" (UID: \"80629dad-7d0a-46b3-bb52-09b75e7dfeb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752358 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/072fbe81-88e1-4755-bb59-5ee95d72760f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752391 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/601c110b-f7ba-4f1c-92ca-dc3cc6634b53-serving-cert\") pod \"authentication-operator-69f744f599-tswct\" (UID: \"601c110b-f7ba-4f1c-92ca-dc3cc6634b53\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752423 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/08e7cd4c-1c51-4142-9384-c0d4561e7e36-signing-key\") pod \"service-ca-9c57cc56f-wdwxg\" (UID: \"08e7cd4c-1c51-4142-9384-c0d4561e7e36\") " pod="openshift-service-ca/service-ca-9c57cc56f-wdwxg" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752459 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpjhl\" (UniqueName: \"kubernetes.io/projected/87c350ba-fd2f-4e8c-a057-be035865dbf4-kube-api-access-bpjhl\") pod \"service-ca-operator-777779d784-7g5t4\" (UID: \"87c350ba-fd2f-4e8c-a057-be035865dbf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7g5t4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752554 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/643cac92-ba9e-4d3c-97b7-ca525f8fc046-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c699m\" (UID: \"643cac92-ba9e-4d3c-97b7-ca525f8fc046\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c699m" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752600 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-serving-cert\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752643 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87c350ba-fd2f-4e8c-a057-be035865dbf4-config\") pod \"service-ca-operator-777779d784-7g5t4\" (UID: \"87c350ba-fd2f-4e8c-a057-be035865dbf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7g5t4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752674 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752707 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/71066f15-4379-475e-8295-671bfd6ea42b-profile-collector-cert\") pod \"catalog-operator-68c6474976-lttkb\" (UID: \"71066f15-4379-475e-8295-671bfd6ea42b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752737 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9j6m\" (UniqueName: \"kubernetes.io/projected/d6db35c3-7baa-4452-b8ff-5c20470142d6-kube-api-access-w9j6m\") pod \"multus-admission-controller-857f4d67dd-dqj92\" (UID: \"d6db35c3-7baa-4452-b8ff-5c20470142d6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dqj92" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752801 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/072fbe81-88e1-4755-bb59-5ee95d72760f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752838 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-audit-policies\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752907 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/71066f15-4379-475e-8295-671bfd6ea42b-srv-cert\") pod \"catalog-operator-68c6474976-lttkb\" (UID: \"71066f15-4379-475e-8295-671bfd6ea42b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.752941 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r8kk\" (UniqueName: \"kubernetes.io/projected/84b7af84-2b2c-48f7-b7b0-e38fb0f7d702-kube-api-access-7r8kk\") pod \"dns-default-kw6xm\" (UID: \"84b7af84-2b2c-48f7-b7b0-e38fb0f7d702\") " pod="openshift-dns/dns-default-kw6xm" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.753064 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/91bd28b3-4940-4027-8753-18335157566e-metrics-tls\") pod \"dns-operator-744455d44c-9t9rg\" (UID: \"91bd28b3-4940-4027-8753-18335157566e\") " pod="openshift-dns-operator/dns-operator-744455d44c-9t9rg" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.753134 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3503ed80-0534-4d0b-a69a-fc9e9c5fa092-profile-collector-cert\") pod \"olm-operator-6b444d44fb-qdgb5\" (UID: \"3503ed80-0534-4d0b-a69a-fc9e9c5fa092\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.753166 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c9bef50a-04f6-427a-813c-bb0039c7aa1a-cert\") pod \"ingress-canary-2dz2z\" (UID: \"c9bef50a-04f6-427a-813c-bb0039c7aa1a\") " pod="openshift-ingress-canary/ingress-canary-2dz2z" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.753270 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/65dfccf6-aa91-41fa-95d6-fd4778065408-ready\") pod \"cni-sysctl-allowlist-ds-dv7nv\" (UID: \"65dfccf6-aa91-41fa-95d6-fd4778065408\") " pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.753355 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/643cac92-ba9e-4d3c-97b7-ca525f8fc046-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c699m\" (UID: \"643cac92-ba9e-4d3c-97b7-ca525f8fc046\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c699m" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.753387 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.753557 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/072fbe81-88e1-4755-bb59-5ee95d72760f-trusted-ca\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.753585 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72d966ad-816d-4483-b9e6-44edb10f9d75-serving-cert\") pod \"route-controller-manager-6576b87f9c-plsdd\" (UID: \"72d966ad-816d-4483-b9e6-44edb10f9d75\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.753662 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-registry-tls\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.753698 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/770461e1-2b0c-4e47-90d1-67f2286e88e2-proxy-tls\") pod \"machine-config-controller-84d6567774-w2rk4\" (UID: \"770461e1-2b0c-4e47-90d1-67f2286e88e2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.753735 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3503ed80-0534-4d0b-a69a-fc9e9c5fa092-srv-cert\") pod \"olm-operator-6b444d44fb-qdgb5\" (UID: \"3503ed80-0534-4d0b-a69a-fc9e9c5fa092\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.754186 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/072fbe81-88e1-4755-bb59-5ee95d72760f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.755317 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3c08c2da-42cd-47ba-92e6-aed1d48855d3-plugins-dir\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.755410 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.755739 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2l6w\" (UniqueName: \"kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-kube-api-access-x2l6w\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.756898 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tslj\" (UniqueName: \"kubernetes.io/projected/739241b1-c772-4ef7-b149-e2098da43d67-kube-api-access-5tslj\") pod \"packageserver-d55dfcdfc-h7bdq\" (UID: \"739241b1-c772-4ef7-b149-e2098da43d67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.757357 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/601c110b-f7ba-4f1c-92ca-dc3cc6634b53-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-tswct\" (UID: \"601c110b-f7ba-4f1c-92ca-dc3cc6634b53\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.759002 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.759098 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w52z\" (UniqueName: \"kubernetes.io/projected/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-kube-api-access-2w52z\") pod \"collect-profiles-29533755-65vjv\" (UID: \"80629dad-7d0a-46b3-bb52-09b75e7dfeb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.759609 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3c08c2da-42cd-47ba-92e6-aed1d48855d3-mountpoint-dir\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.759704 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.760030 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-bound-sa-token\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.760335 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txx4q\" (UniqueName: \"kubernetes.io/projected/6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c-kube-api-access-txx4q\") pod \"machine-config-server-ccjcb\" (UID: \"6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c\") " pod="openshift-machine-config-operator/machine-config-server-ccjcb" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.760412 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a14e9f5d-f4ac-436b-8a6e-9523697d2b74-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-shdc4\" (UID: \"a14e9f5d-f4ac-436b-8a6e-9523697d2b74\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-shdc4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.760818 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsqqc\" (UniqueName: \"kubernetes.io/projected/a14e9f5d-f4ac-436b-8a6e-9523697d2b74-kube-api-access-rsqqc\") pod \"cluster-samples-operator-665b6dd947-shdc4\" (UID: \"a14e9f5d-f4ac-436b-8a6e-9523697d2b74\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-shdc4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.760916 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/65dfccf6-aa91-41fa-95d6-fd4778065408-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-dv7nv\" (UID: \"65dfccf6-aa91-41fa-95d6-fd4778065408\") " pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.762645 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c-certs\") pod \"machine-config-server-ccjcb\" (UID: \"6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c\") " pod="openshift-machine-config-operator/machine-config-server-ccjcb" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.762743 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/739241b1-c772-4ef7-b149-e2098da43d67-apiservice-cert\") pod \"packageserver-d55dfcdfc-h7bdq\" (UID: \"739241b1-c772-4ef7-b149-e2098da43d67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.762928 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/72d966ad-816d-4483-b9e6-44edb10f9d75-client-ca\") pod \"route-controller-manager-6576b87f9c-plsdd\" (UID: \"72d966ad-816d-4483-b9e6-44edb10f9d75\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.763023 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.763238 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ebefeb1c-aa64-4a51-8d2d-197d8e8421af-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jcf7k\" (UID: \"ebefeb1c-aa64-4a51-8d2d-197d8e8421af\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jcf7k" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.763333 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-oauth-serving-cert\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.763486 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c-node-bootstrap-token\") pod \"machine-config-server-ccjcb\" (UID: \"6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c\") " pod="openshift-machine-config-operator/machine-config-server-ccjcb" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.763749 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw7jm\" (UniqueName: \"kubernetes.io/projected/08e7cd4c-1c51-4142-9384-c0d4561e7e36-kube-api-access-fw7jm\") pod \"service-ca-9c57cc56f-wdwxg\" (UID: \"08e7cd4c-1c51-4142-9384-c0d4561e7e36\") " pod="openshift-service-ca/service-ca-9c57cc56f-wdwxg" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.763952 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxkh9\" (UniqueName: \"kubernetes.io/projected/c9bef50a-04f6-427a-813c-bb0039c7aa1a-kube-api-access-zxkh9\") pod \"ingress-canary-2dz2z\" (UID: \"c9bef50a-04f6-427a-813c-bb0039c7aa1a\") " pod="openshift-ingress-canary/ingress-canary-2dz2z" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.764034 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdc0843c-6110-4d69-b444-7e5384f858e1-serving-cert\") pod \"console-operator-58897d9998-w8lg6\" (UID: \"bdc0843c-6110-4d69-b444-7e5384f858e1\") " pod="openshift-console-operator/console-operator-58897d9998-w8lg6" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.764160 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b2fb1a6-e1d1-48e0-83dc-17143971294e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6th6z\" (UID: \"1b2fb1a6-e1d1-48e0-83dc-17143971294e\") " pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.764317 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.764555 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/887d2231-1877-4977-ac7d-9f7cf1b33a23-audit-dir\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.764702 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/84b7af84-2b2c-48f7-b7b0-e38fb0f7d702-metrics-tls\") pod \"dns-default-kw6xm\" (UID: \"84b7af84-2b2c-48f7-b7b0-e38fb0f7d702\") " pod="openshift-dns/dns-default-kw6xm" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.764802 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/601c110b-f7ba-4f1c-92ca-dc3cc6634b53-config\") pod \"authentication-operator-69f744f599-tswct\" (UID: \"601c110b-f7ba-4f1c-92ca-dc3cc6634b53\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.764913 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.765068 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-secret-volume\") pod \"collect-profiles-29533755-65vjv\" (UID: \"80629dad-7d0a-46b3-bb52-09b75e7dfeb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.765164 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/072fbe81-88e1-4755-bb59-5ee95d72760f-registry-certificates\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.765234 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3c08c2da-42cd-47ba-92e6-aed1d48855d3-socket-dir\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.765459 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/072fbe81-88e1-4755-bb59-5ee95d72760f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.766207 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.768376 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/072fbe81-88e1-4755-bb59-5ee95d72760f-registry-certificates\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.768753 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-registry-tls\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.771914 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.785383 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.806295 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.825311 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.846062 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.865976 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.866573 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.866779 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: E0225 13:22:31.866823 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:32.366779922 +0000 UTC m=+110.167878006 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.866891 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b2fb1a6-e1d1-48e0-83dc-17143971294e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6th6z\" (UID: \"1b2fb1a6-e1d1-48e0-83dc-17143971294e\") " pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.866991 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/887d2231-1877-4977-ac7d-9f7cf1b33a23-audit-dir\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.867064 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/84b7af84-2b2c-48f7-b7b0-e38fb0f7d702-metrics-tls\") pod \"dns-default-kw6xm\" (UID: \"84b7af84-2b2c-48f7-b7b0-e38fb0f7d702\") " pod="openshift-dns/dns-default-kw6xm" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.867114 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/601c110b-f7ba-4f1c-92ca-dc3cc6634b53-config\") pod \"authentication-operator-69f744f599-tswct\" (UID: \"601c110b-f7ba-4f1c-92ca-dc3cc6634b53\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.867166 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.867175 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/887d2231-1877-4977-ac7d-9f7cf1b33a23-audit-dir\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.867232 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-secret-volume\") pod \"collect-profiles-29533755-65vjv\" (UID: \"80629dad-7d0a-46b3-bb52-09b75e7dfeb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.867275 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3c08c2da-42cd-47ba-92e6-aed1d48855d3-socket-dir\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.868135 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3c08c2da-42cd-47ba-92e6-aed1d48855d3-csi-data-dir\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.868229 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.868288 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/739241b1-c772-4ef7-b149-e2098da43d67-webhook-cert\") pod \"packageserver-d55dfcdfc-h7bdq\" (UID: \"739241b1-c772-4ef7-b149-e2098da43d67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.868619 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-encryption-config\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.868729 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dv4r\" (UniqueName: \"kubernetes.io/projected/bdc0843c-6110-4d69-b444-7e5384f858e1-kube-api-access-7dv4r\") pod \"console-operator-58897d9998-w8lg6\" (UID: \"bdc0843c-6110-4d69-b444-7e5384f858e1\") " pod="openshift-console-operator/console-operator-58897d9998-w8lg6" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.868863 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7rzd\" (UniqueName: \"kubernetes.io/projected/71066f15-4379-475e-8295-671bfd6ea42b-kube-api-access-q7rzd\") pod \"catalog-operator-68c6474976-lttkb\" (UID: \"71066f15-4379-475e-8295-671bfd6ea42b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.868920 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/601c110b-f7ba-4f1c-92ca-dc3cc6634b53-config\") pod \"authentication-operator-69f744f599-tswct\" (UID: \"601c110b-f7ba-4f1c-92ca-dc3cc6634b53\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.868421 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3c08c2da-42cd-47ba-92e6-aed1d48855d3-socket-dir\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.869089 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-etcd-client\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.868466 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3c08c2da-42cd-47ba-92e6-aed1d48855d3-csi-data-dir\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.869216 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86tlk\" (UniqueName: \"kubernetes.io/projected/887d2231-1877-4977-ac7d-9f7cf1b33a23-kube-api-access-86tlk\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.869326 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.869390 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-oauth-config\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.869438 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72d966ad-816d-4483-b9e6-44edb10f9d75-config\") pod \"route-controller-manager-6576b87f9c-plsdd\" (UID: \"72d966ad-816d-4483-b9e6-44edb10f9d75\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.869552 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.869615 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlj7n\" (UniqueName: \"kubernetes.io/projected/ebefeb1c-aa64-4a51-8d2d-197d8e8421af-kube-api-access-mlj7n\") pod \"openshift-controller-manager-operator-756b6f6bc6-jcf7k\" (UID: \"ebefeb1c-aa64-4a51-8d2d-197d8e8421af\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jcf7k" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.869663 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-service-ca\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.869699 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfh4w\" (UniqueName: \"kubernetes.io/projected/3c08c2da-42cd-47ba-92e6-aed1d48855d3-kube-api-access-lfh4w\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.869735 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/770461e1-2b0c-4e47-90d1-67f2286e88e2-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-w2rk4\" (UID: \"770461e1-2b0c-4e47-90d1-67f2286e88e2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.869797 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp9x9\" (UniqueName: \"kubernetes.io/projected/601c110b-f7ba-4f1c-92ca-dc3cc6634b53-kube-api-access-zp9x9\") pod \"authentication-operator-69f744f599-tswct\" (UID: \"601c110b-f7ba-4f1c-92ca-dc3cc6634b53\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.869838 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-audit-policies\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.869872 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.869909 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs2dt\" (UniqueName: \"kubernetes.io/projected/3503ed80-0534-4d0b-a69a-fc9e9c5fa092-kube-api-access-fs2dt\") pod \"olm-operator-6b444d44fb-qdgb5\" (UID: \"3503ed80-0534-4d0b-a69a-fc9e9c5fa092\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.869948 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmrn9\" (UniqueName: \"kubernetes.io/projected/91bd28b3-4940-4027-8753-18335157566e-kube-api-access-vmrn9\") pod \"dns-operator-744455d44c-9t9rg\" (UID: \"91bd28b3-4940-4027-8753-18335157566e\") " pod="openshift-dns-operator/dns-operator-744455d44c-9t9rg" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.869984 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87c350ba-fd2f-4e8c-a057-be035865dbf4-serving-cert\") pod \"service-ca-operator-777779d784-7g5t4\" (UID: \"87c350ba-fd2f-4e8c-a057-be035865dbf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7g5t4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870032 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/643cac92-ba9e-4d3c-97b7-ca525f8fc046-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c699m\" (UID: \"643cac92-ba9e-4d3c-97b7-ca525f8fc046\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c699m" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870068 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870104 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k82t\" (UniqueName: \"kubernetes.io/projected/6403af09-14ef-49fc-858f-06c1bbadb88b-kube-api-access-7k82t\") pod \"auto-csr-approver-29533762-lkzv9\" (UID: \"6403af09-14ef-49fc-858f-06c1bbadb88b\") " pod="openshift-infra/auto-csr-approver-29533762-lkzv9" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870142 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6zdr\" (UniqueName: \"kubernetes.io/projected/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-kube-api-access-c6zdr\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870181 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggzb8\" (UniqueName: \"kubernetes.io/projected/65dfccf6-aa91-41fa-95d6-fd4778065408-kube-api-access-ggzb8\") pod \"cni-sysctl-allowlist-ds-dv7nv\" (UID: \"65dfccf6-aa91-41fa-95d6-fd4778065408\") " pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870222 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdc0843c-6110-4d69-b444-7e5384f858e1-config\") pod \"console-operator-58897d9998-w8lg6\" (UID: \"bdc0843c-6110-4d69-b444-7e5384f858e1\") " pod="openshift-console-operator/console-operator-58897d9998-w8lg6" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870258 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr9jx\" (UniqueName: \"kubernetes.io/projected/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-kube-api-access-hr9jx\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870302 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-trusted-ca-bundle\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870346 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d6db35c3-7baa-4452-b8ff-5c20470142d6-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dqj92\" (UID: \"d6db35c3-7baa-4452-b8ff-5c20470142d6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dqj92" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870381 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/739241b1-c772-4ef7-b149-e2098da43d67-tmpfs\") pod \"packageserver-d55dfcdfc-h7bdq\" (UID: \"739241b1-c772-4ef7-b149-e2098da43d67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870420 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brbmt\" (UniqueName: \"kubernetes.io/projected/72d966ad-816d-4483-b9e6-44edb10f9d75-kube-api-access-brbmt\") pod \"route-controller-manager-6576b87f9c-plsdd\" (UID: \"72d966ad-816d-4483-b9e6-44edb10f9d75\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870457 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/08e7cd4c-1c51-4142-9384-c0d4561e7e36-signing-cabundle\") pod \"service-ca-9c57cc56f-wdwxg\" (UID: \"08e7cd4c-1c51-4142-9384-c0d4561e7e36\") " pod="openshift-service-ca/service-ca-9c57cc56f-wdwxg" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870494 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bdc0843c-6110-4d69-b444-7e5384f858e1-trusted-ca\") pod \"console-operator-58897d9998-w8lg6\" (UID: \"bdc0843c-6110-4d69-b444-7e5384f858e1\") " pod="openshift-console-operator/console-operator-58897d9998-w8lg6" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870558 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksc96\" (UniqueName: \"kubernetes.io/projected/7bcf8b34-940d-41d9-b32a-5548c5cc62c2-kube-api-access-ksc96\") pod \"package-server-manager-789f6589d5-njtsc\" (UID: \"7bcf8b34-940d-41d9-b32a-5548c5cc62c2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-njtsc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870595 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870634 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km5v7\" (UniqueName: \"kubernetes.io/projected/1b2fb1a6-e1d1-48e0-83dc-17143971294e-kube-api-access-km5v7\") pod \"marketplace-operator-79b997595-6th6z\" (UID: \"1b2fb1a6-e1d1-48e0-83dc-17143971294e\") " pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870669 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-audit-dir\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870704 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/65dfccf6-aa91-41fa-95d6-fd4778065408-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-dv7nv\" (UID: \"65dfccf6-aa91-41fa-95d6-fd4778065408\") " pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870741 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3c08c2da-42cd-47ba-92e6-aed1d48855d3-registration-dir\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: E0225 13:22:31.870783 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:32.370763924 +0000 UTC m=+110.171862008 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870829 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdvqb\" (UniqueName: \"kubernetes.io/projected/770461e1-2b0c-4e47-90d1-67f2286e88e2-kube-api-access-qdvqb\") pod \"machine-config-controller-84d6567774-w2rk4\" (UID: \"770461e1-2b0c-4e47-90d1-67f2286e88e2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870871 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3c08c2da-42cd-47ba-92e6-aed1d48855d3-registration-dir\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870892 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebefeb1c-aa64-4a51-8d2d-197d8e8421af-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jcf7k\" (UID: \"ebefeb1c-aa64-4a51-8d2d-197d8e8421af\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jcf7k" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870927 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-serving-cert\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.870961 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-config\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871016 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84b7af84-2b2c-48f7-b7b0-e38fb0f7d702-config-volume\") pod \"dns-default-kw6xm\" (UID: \"84b7af84-2b2c-48f7-b7b0-e38fb0f7d702\") " pod="openshift-dns/dns-default-kw6xm" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871070 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bcf8b34-940d-41d9-b32a-5548c5cc62c2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-njtsc\" (UID: \"7bcf8b34-940d-41d9-b32a-5548c5cc62c2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-njtsc" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871103 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871145 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1b2fb1a6-e1d1-48e0-83dc-17143971294e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6th6z\" (UID: \"1b2fb1a6-e1d1-48e0-83dc-17143971294e\") " pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871181 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/601c110b-f7ba-4f1c-92ca-dc3cc6634b53-service-ca-bundle\") pod \"authentication-operator-69f744f599-tswct\" (UID: \"601c110b-f7ba-4f1c-92ca-dc3cc6634b53\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871212 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-config-volume\") pod \"collect-profiles-29533755-65vjv\" (UID: \"80629dad-7d0a-46b3-bb52-09b75e7dfeb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871268 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/601c110b-f7ba-4f1c-92ca-dc3cc6634b53-serving-cert\") pod \"authentication-operator-69f744f599-tswct\" (UID: \"601c110b-f7ba-4f1c-92ca-dc3cc6634b53\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871303 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpjhl\" (UniqueName: \"kubernetes.io/projected/87c350ba-fd2f-4e8c-a057-be035865dbf4-kube-api-access-bpjhl\") pod \"service-ca-operator-777779d784-7g5t4\" (UID: \"87c350ba-fd2f-4e8c-a057-be035865dbf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7g5t4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871336 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/08e7cd4c-1c51-4142-9384-c0d4561e7e36-signing-key\") pod \"service-ca-9c57cc56f-wdwxg\" (UID: \"08e7cd4c-1c51-4142-9384-c0d4561e7e36\") " pod="openshift-service-ca/service-ca-9c57cc56f-wdwxg" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871369 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/643cac92-ba9e-4d3c-97b7-ca525f8fc046-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c699m\" (UID: \"643cac92-ba9e-4d3c-97b7-ca525f8fc046\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c699m" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871404 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-serving-cert\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871444 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871480 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87c350ba-fd2f-4e8c-a057-be035865dbf4-config\") pod \"service-ca-operator-777779d784-7g5t4\" (UID: \"87c350ba-fd2f-4e8c-a057-be035865dbf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7g5t4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871558 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/71066f15-4379-475e-8295-671bfd6ea42b-profile-collector-cert\") pod \"catalog-operator-68c6474976-lttkb\" (UID: \"71066f15-4379-475e-8295-671bfd6ea42b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871609 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9j6m\" (UniqueName: \"kubernetes.io/projected/d6db35c3-7baa-4452-b8ff-5c20470142d6-kube-api-access-w9j6m\") pod \"multus-admission-controller-857f4d67dd-dqj92\" (UID: \"d6db35c3-7baa-4452-b8ff-5c20470142d6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dqj92" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871659 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-audit-policies\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871683 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72d966ad-816d-4483-b9e6-44edb10f9d75-config\") pod \"route-controller-manager-6576b87f9c-plsdd\" (UID: \"72d966ad-816d-4483-b9e6-44edb10f9d75\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871728 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/71066f15-4379-475e-8295-671bfd6ea42b-srv-cert\") pod \"catalog-operator-68c6474976-lttkb\" (UID: \"71066f15-4379-475e-8295-671bfd6ea42b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871785 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r8kk\" (UniqueName: \"kubernetes.io/projected/84b7af84-2b2c-48f7-b7b0-e38fb0f7d702-kube-api-access-7r8kk\") pod \"dns-default-kw6xm\" (UID: \"84b7af84-2b2c-48f7-b7b0-e38fb0f7d702\") " pod="openshift-dns/dns-default-kw6xm" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871856 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/91bd28b3-4940-4027-8753-18335157566e-metrics-tls\") pod \"dns-operator-744455d44c-9t9rg\" (UID: \"91bd28b3-4940-4027-8753-18335157566e\") " pod="openshift-dns-operator/dns-operator-744455d44c-9t9rg" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871904 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3503ed80-0534-4d0b-a69a-fc9e9c5fa092-profile-collector-cert\") pod \"olm-operator-6b444d44fb-qdgb5\" (UID: \"3503ed80-0534-4d0b-a69a-fc9e9c5fa092\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871951 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c9bef50a-04f6-427a-813c-bb0039c7aa1a-cert\") pod \"ingress-canary-2dz2z\" (UID: \"c9bef50a-04f6-427a-813c-bb0039c7aa1a\") " pod="openshift-ingress-canary/ingress-canary-2dz2z" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.871998 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/65dfccf6-aa91-41fa-95d6-fd4778065408-ready\") pod \"cni-sysctl-allowlist-ds-dv7nv\" (UID: \"65dfccf6-aa91-41fa-95d6-fd4778065408\") " pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872054 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72d966ad-816d-4483-b9e6-44edb10f9d75-serving-cert\") pod \"route-controller-manager-6576b87f9c-plsdd\" (UID: \"72d966ad-816d-4483-b9e6-44edb10f9d75\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872077 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/739241b1-c772-4ef7-b149-e2098da43d67-tmpfs\") pod \"packageserver-d55dfcdfc-h7bdq\" (UID: \"739241b1-c772-4ef7-b149-e2098da43d67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872091 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/643cac92-ba9e-4d3c-97b7-ca525f8fc046-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c699m\" (UID: \"643cac92-ba9e-4d3c-97b7-ca525f8fc046\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c699m" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872129 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872167 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/770461e1-2b0c-4e47-90d1-67f2286e88e2-proxy-tls\") pod \"machine-config-controller-84d6567774-w2rk4\" (UID: \"770461e1-2b0c-4e47-90d1-67f2286e88e2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872202 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3503ed80-0534-4d0b-a69a-fc9e9c5fa092-srv-cert\") pod \"olm-operator-6b444d44fb-qdgb5\" (UID: \"3503ed80-0534-4d0b-a69a-fc9e9c5fa092\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872294 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3c08c2da-42cd-47ba-92e6-aed1d48855d3-plugins-dir\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872331 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872376 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tslj\" (UniqueName: \"kubernetes.io/projected/739241b1-c772-4ef7-b149-e2098da43d67-kube-api-access-5tslj\") pod \"packageserver-d55dfcdfc-h7bdq\" (UID: \"739241b1-c772-4ef7-b149-e2098da43d67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872430 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/601c110b-f7ba-4f1c-92ca-dc3cc6634b53-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-tswct\" (UID: \"601c110b-f7ba-4f1c-92ca-dc3cc6634b53\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872483 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872546 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w52z\" (UniqueName: \"kubernetes.io/projected/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-kube-api-access-2w52z\") pod \"collect-profiles-29533755-65vjv\" (UID: \"80629dad-7d0a-46b3-bb52-09b75e7dfeb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872584 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3c08c2da-42cd-47ba-92e6-aed1d48855d3-mountpoint-dir\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872591 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/739241b1-c772-4ef7-b149-e2098da43d67-webhook-cert\") pod \"packageserver-d55dfcdfc-h7bdq\" (UID: \"739241b1-c772-4ef7-b149-e2098da43d67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872618 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872722 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txx4q\" (UniqueName: \"kubernetes.io/projected/6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c-kube-api-access-txx4q\") pod \"machine-config-server-ccjcb\" (UID: \"6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c\") " pod="openshift-machine-config-operator/machine-config-server-ccjcb" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872777 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a14e9f5d-f4ac-436b-8a6e-9523697d2b74-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-shdc4\" (UID: \"a14e9f5d-f4ac-436b-8a6e-9523697d2b74\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-shdc4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872818 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsqqc\" (UniqueName: \"kubernetes.io/projected/a14e9f5d-f4ac-436b-8a6e-9523697d2b74-kube-api-access-rsqqc\") pod \"cluster-samples-operator-665b6dd947-shdc4\" (UID: \"a14e9f5d-f4ac-436b-8a6e-9523697d2b74\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-shdc4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872854 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/65dfccf6-aa91-41fa-95d6-fd4778065408-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-dv7nv\" (UID: \"65dfccf6-aa91-41fa-95d6-fd4778065408\") " pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872917 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/739241b1-c772-4ef7-b149-e2098da43d67-apiservice-cert\") pod \"packageserver-d55dfcdfc-h7bdq\" (UID: \"739241b1-c772-4ef7-b149-e2098da43d67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872952 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c-certs\") pod \"machine-config-server-ccjcb\" (UID: \"6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c\") " pod="openshift-machine-config-operator/machine-config-server-ccjcb" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.872990 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/72d966ad-816d-4483-b9e6-44edb10f9d75-client-ca\") pod \"route-controller-manager-6576b87f9c-plsdd\" (UID: \"72d966ad-816d-4483-b9e6-44edb10f9d75\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.873028 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.873103 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-oauth-serving-cert\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.873146 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ebefeb1c-aa64-4a51-8d2d-197d8e8421af-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jcf7k\" (UID: \"ebefeb1c-aa64-4a51-8d2d-197d8e8421af\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jcf7k" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.873158 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-service-ca\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.873190 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c-node-bootstrap-token\") pod \"machine-config-server-ccjcb\" (UID: \"6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c\") " pod="openshift-machine-config-operator/machine-config-server-ccjcb" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.873233 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw7jm\" (UniqueName: \"kubernetes.io/projected/08e7cd4c-1c51-4142-9384-c0d4561e7e36-kube-api-access-fw7jm\") pod \"service-ca-9c57cc56f-wdwxg\" (UID: \"08e7cd4c-1c51-4142-9384-c0d4561e7e36\") " pod="openshift-service-ca/service-ca-9c57cc56f-wdwxg" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.873271 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxkh9\" (UniqueName: \"kubernetes.io/projected/c9bef50a-04f6-427a-813c-bb0039c7aa1a-kube-api-access-zxkh9\") pod \"ingress-canary-2dz2z\" (UID: \"c9bef50a-04f6-427a-813c-bb0039c7aa1a\") " pod="openshift-ingress-canary/ingress-canary-2dz2z" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.873309 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdc0843c-6110-4d69-b444-7e5384f858e1-serving-cert\") pod \"console-operator-58897d9998-w8lg6\" (UID: \"bdc0843c-6110-4d69-b444-7e5384f858e1\") " pod="openshift-console-operator/console-operator-58897d9998-w8lg6" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.873546 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-trusted-ca-bundle\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.873902 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/65dfccf6-aa91-41fa-95d6-fd4778065408-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-dv7nv\" (UID: \"65dfccf6-aa91-41fa-95d6-fd4778065408\") " pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.874455 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/643cac92-ba9e-4d3c-97b7-ca525f8fc046-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c699m\" (UID: \"643cac92-ba9e-4d3c-97b7-ca525f8fc046\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c699m" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.874651 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-audit-policies\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.874977 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-encryption-config\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.875720 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-oauth-serving-cert\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.875759 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-audit-dir\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.876668 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/770461e1-2b0c-4e47-90d1-67f2286e88e2-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-w2rk4\" (UID: \"770461e1-2b0c-4e47-90d1-67f2286e88e2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.877556 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebefeb1c-aa64-4a51-8d2d-197d8e8421af-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jcf7k\" (UID: \"ebefeb1c-aa64-4a51-8d2d-197d8e8421af\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jcf7k" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.877950 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.879901 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3c08c2da-42cd-47ba-92e6-aed1d48855d3-plugins-dir\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.880079 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3c08c2da-42cd-47ba-92e6-aed1d48855d3-mountpoint-dir\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.881479 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d6db35c3-7baa-4452-b8ff-5c20470142d6-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dqj92\" (UID: \"d6db35c3-7baa-4452-b8ff-5c20470142d6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dqj92" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.881982 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ebefeb1c-aa64-4a51-8d2d-197d8e8421af-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jcf7k\" (UID: \"ebefeb1c-aa64-4a51-8d2d-197d8e8421af\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jcf7k" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.882662 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-etcd-client\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.883126 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/65dfccf6-aa91-41fa-95d6-fd4778065408-ready\") pod \"cni-sysctl-allowlist-ds-dv7nv\" (UID: \"65dfccf6-aa91-41fa-95d6-fd4778065408\") " pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.883165 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-config\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.883281 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.883734 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/601c110b-f7ba-4f1c-92ca-dc3cc6634b53-service-ca-bundle\") pod \"authentication-operator-69f744f599-tswct\" (UID: \"601c110b-f7ba-4f1c-92ca-dc3cc6634b53\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.883875 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-oauth-config\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.884045 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3503ed80-0534-4d0b-a69a-fc9e9c5fa092-profile-collector-cert\") pod \"olm-operator-6b444d44fb-qdgb5\" (UID: \"3503ed80-0534-4d0b-a69a-fc9e9c5fa092\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.884729 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.885468 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.885731 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.885732 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-serving-cert\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.886127 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/601c110b-f7ba-4f1c-92ca-dc3cc6634b53-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-tswct\" (UID: \"601c110b-f7ba-4f1c-92ca-dc3cc6634b53\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.886650 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.886764 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-secret-volume\") pod \"collect-profiles-29533755-65vjv\" (UID: \"80629dad-7d0a-46b3-bb52-09b75e7dfeb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.888011 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3503ed80-0534-4d0b-a69a-fc9e9c5fa092-srv-cert\") pod \"olm-operator-6b444d44fb-qdgb5\" (UID: \"3503ed80-0534-4d0b-a69a-fc9e9c5fa092\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.888459 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/72d966ad-816d-4483-b9e6-44edb10f9d75-client-ca\") pod \"route-controller-manager-6576b87f9c-plsdd\" (UID: \"72d966ad-816d-4483-b9e6-44edb10f9d75\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.888543 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.889092 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.889122 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.889381 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/739241b1-c772-4ef7-b149-e2098da43d67-apiservice-cert\") pod \"packageserver-d55dfcdfc-h7bdq\" (UID: \"739241b1-c772-4ef7-b149-e2098da43d67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.889917 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.890086 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/643cac92-ba9e-4d3c-97b7-ca525f8fc046-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c699m\" (UID: \"643cac92-ba9e-4d3c-97b7-ca525f8fc046\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c699m" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.890096 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/770461e1-2b0c-4e47-90d1-67f2286e88e2-proxy-tls\") pod \"machine-config-controller-84d6567774-w2rk4\" (UID: \"770461e1-2b0c-4e47-90d1-67f2286e88e2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.890242 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.890550 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.890914 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/71066f15-4379-475e-8295-671bfd6ea42b-profile-collector-cert\") pod \"catalog-operator-68c6474976-lttkb\" (UID: \"71066f15-4379-475e-8295-671bfd6ea42b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.891280 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/601c110b-f7ba-4f1c-92ca-dc3cc6634b53-serving-cert\") pod \"authentication-operator-69f744f599-tswct\" (UID: \"601c110b-f7ba-4f1c-92ca-dc3cc6634b53\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.891425 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/71066f15-4379-475e-8295-671bfd6ea42b-srv-cert\") pod \"catalog-operator-68c6474976-lttkb\" (UID: \"71066f15-4379-475e-8295-671bfd6ea42b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.892717 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72d966ad-816d-4483-b9e6-44edb10f9d75-serving-cert\") pod \"route-controller-manager-6576b87f9c-plsdd\" (UID: \"72d966ad-816d-4483-b9e6-44edb10f9d75\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.893065 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.893537 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-serving-cert\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.925895 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.934871 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.935001 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.934892 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.943813 4815 request.go:700] Waited for 1.001505772s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-multus/secrets?fieldSelector=metadata.name%3Dmetrics-daemon-sa-dockercfg-d427c&limit=500&resourceVersion=0 Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.946076 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.965006 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.968898 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-audit-policies\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.973949 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:31 crc kubenswrapper[4815]: E0225 13:22:31.974036 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:32.474019039 +0000 UTC m=+110.275117093 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.974491 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:31 crc kubenswrapper[4815]: E0225 13:22:31.974898 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:32.474882475 +0000 UTC m=+110.275980529 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.985158 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 25 13:22:31 crc kubenswrapper[4815]: I0225 13:22:31.995785 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bcf8b34-940d-41d9-b32a-5548c5cc62c2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-njtsc\" (UID: \"7bcf8b34-940d-41d9-b32a-5548c5cc62c2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-njtsc" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.005395 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.011271 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1b2fb1a6-e1d1-48e0-83dc-17143971294e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6th6z\" (UID: \"1b2fb1a6-e1d1-48e0-83dc-17143971294e\") " pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.025083 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.052945 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.060174 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b2fb1a6-e1d1-48e0-83dc-17143971294e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6th6z\" (UID: \"1b2fb1a6-e1d1-48e0-83dc-17143971294e\") " pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.064689 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.076040 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.076761 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:32.576746258 +0000 UTC m=+110.377844302 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.085997 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.105271 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.110919 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-config-volume\") pod \"collect-profiles-29533755-65vjv\" (UID: \"80629dad-7d0a-46b3-bb52-09b75e7dfeb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.125620 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.146838 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.166228 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.177377 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.177949 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:32.677923871 +0000 UTC m=+110.479021965 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.185234 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.192840 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/08e7cd4c-1c51-4142-9384-c0d4561e7e36-signing-key\") pod \"service-ca-9c57cc56f-wdwxg\" (UID: \"08e7cd4c-1c51-4142-9384-c0d4561e7e36\") " pod="openshift-service-ca/service-ca-9c57cc56f-wdwxg" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.206052 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.225793 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.233822 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/08e7cd4c-1c51-4142-9384-c0d4561e7e36-signing-cabundle\") pod \"service-ca-9c57cc56f-wdwxg\" (UID: \"08e7cd4c-1c51-4142-9384-c0d4561e7e36\") " pod="openshift-service-ca/service-ca-9c57cc56f-wdwxg" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.246144 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.265421 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.279438 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.279797 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:32.779767374 +0000 UTC m=+110.580865468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.280772 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.281188 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:32.781168296 +0000 UTC m=+110.582266470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.285765 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.298021 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a14e9f5d-f4ac-436b-8a6e-9523697d2b74-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-shdc4\" (UID: \"a14e9f5d-f4ac-436b-8a6e-9523697d2b74\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-shdc4" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.305871 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.325313 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.345369 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.365033 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.382361 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.382463 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:32.882444371 +0000 UTC m=+110.683542435 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.382786 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.383270 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:32.883248556 +0000 UTC m=+110.684346620 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.386084 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.406002 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.416277 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdc0843c-6110-4d69-b444-7e5384f858e1-config\") pod \"console-operator-58897d9998-w8lg6\" (UID: \"bdc0843c-6110-4d69-b444-7e5384f858e1\") " pod="openshift-console-operator/console-operator-58897d9998-w8lg6" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.437828 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.444386 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.444496 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bdc0843c-6110-4d69-b444-7e5384f858e1-trusted-ca\") pod \"console-operator-58897d9998-w8lg6\" (UID: \"bdc0843c-6110-4d69-b444-7e5384f858e1\") " pod="openshift-console-operator/console-operator-58897d9998-w8lg6" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.459856 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdc0843c-6110-4d69-b444-7e5384f858e1-serving-cert\") pod \"console-operator-58897d9998-w8lg6\" (UID: \"bdc0843c-6110-4d69-b444-7e5384f858e1\") " pod="openshift-console-operator/console-operator-58897d9998-w8lg6" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.464983 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.484653 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.484917 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:32.984891432 +0000 UTC m=+110.785989516 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.485547 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.485971 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:32.985943004 +0000 UTC m=+110.787041148 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.486156 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.490464 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87c350ba-fd2f-4e8c-a057-be035865dbf4-config\") pod \"service-ca-operator-777779d784-7g5t4\" (UID: \"87c350ba-fd2f-4e8c-a057-be035865dbf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7g5t4" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.505501 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.525811 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.536719 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87c350ba-fd2f-4e8c-a057-be035865dbf4-serving-cert\") pod \"service-ca-operator-777779d784-7g5t4\" (UID: \"87c350ba-fd2f-4e8c-a057-be035865dbf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7g5t4" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.546875 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.565646 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.585782 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.586224 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.586369 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:33.086341543 +0000 UTC m=+110.887439617 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.586565 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.586893 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:33.086879229 +0000 UTC m=+110.887977293 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.605489 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.612045 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/91bd28b3-4940-4027-8753-18335157566e-metrics-tls\") pod \"dns-operator-744455d44c-9t9rg\" (UID: \"91bd28b3-4940-4027-8753-18335157566e\") " pod="openshift-dns-operator/dns-operator-744455d44c-9t9rg" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.625458 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.644798 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.665632 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-sysctl-allowlist" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.667077 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/65dfccf6-aa91-41fa-95d6-fd4778065408-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-dv7nv\" (UID: \"65dfccf6-aa91-41fa-95d6-fd4778065408\") " pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.684624 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.688132 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.688392 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:33.188359511 +0000 UTC m=+110.989457625 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.688730 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84b7af84-2b2c-48f7-b7b0-e38fb0f7d702-config-volume\") pod \"dns-default-kw6xm\" (UID: \"84b7af84-2b2c-48f7-b7b0-e38fb0f7d702\") " pod="openshift-dns/dns-default-kw6xm" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.689373 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.689918 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:33.189901037 +0000 UTC m=+110.990999121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.705796 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.725207 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.734079 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/84b7af84-2b2c-48f7-b7b0-e38fb0f7d702-metrics-tls\") pod \"dns-default-kw6xm\" (UID: \"84b7af84-2b2c-48f7-b7b0-e38fb0f7d702\") " pod="openshift-dns/dns-default-kw6xm" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.745827 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.762852 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c-certs\") pod \"machine-config-server-ccjcb\" (UID: \"6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c\") " pod="openshift-machine-config-operator/machine-config-server-ccjcb" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.765299 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.779229 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c-node-bootstrap-token\") pod \"machine-config-server-ccjcb\" (UID: \"6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c\") " pod="openshift-machine-config-operator/machine-config-server-ccjcb" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.786244 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.790866 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.790984 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:33.290958126 +0000 UTC m=+111.092056200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.791706 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.792206 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:33.292182914 +0000 UTC m=+111.093280998 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.805903 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.825573 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.833437 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c9bef50a-04f6-427a-813c-bb0039c7aa1a-cert\") pod \"ingress-canary-2dz2z\" (UID: \"c9bef50a-04f6-427a-813c-bb0039c7aa1a\") " pod="openshift-ingress-canary/ingress-canary-2dz2z" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.845909 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.865559 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.885706 4815 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.892844 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.893008 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:33.392981474 +0000 UTC m=+111.194079568 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.893700 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.894303 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:33.394280413 +0000 UTC m=+111.195378507 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.905590 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.925281 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.963599 4815 request.go:700] Waited for 1.925981834s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/serviceaccounts/kube-storage-version-migrator-operator/token Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.972413 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d5sq\" (UniqueName: \"kubernetes.io/projected/2a76074e-1587-4795-a538-73d1fee5b028-kube-api-access-2d5sq\") pod \"cluster-image-registry-operator-dc59b4c8b-6pxmc\" (UID: \"2a76074e-1587-4795-a538-73d1fee5b028\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.992001 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4828s\" (UniqueName: \"kubernetes.io/projected/a2f6b143-248f-4851-966e-93009db4876b-kube-api-access-4828s\") pod \"kube-storage-version-migrator-operator-b67b599dd-qscdn\" (UID: \"a2f6b143-248f-4851-966e-93009db4876b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qscdn" Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.995899 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.996143 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:33.496112276 +0000 UTC m=+111.297210370 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:32 crc kubenswrapper[4815]: I0225 13:22:32.996395 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:32 crc kubenswrapper[4815]: E0225 13:22:32.996884 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:33.496867758 +0000 UTC m=+111.297965852 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.005024 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp5w9\" (UniqueName: \"kubernetes.io/projected/78d30b59-0a66-44e5-b0b9-a2864e4d4552-kube-api-access-lp5w9\") pod \"downloads-7954f5f757-8jnnl\" (UID: \"78d30b59-0a66-44e5-b0b9-a2864e4d4552\") " pod="openshift-console/downloads-7954f5f757-8jnnl" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.032084 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qscdn" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.033323 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s59t\" (UniqueName: \"kubernetes.io/projected/6f28beca-7506-4390-9c74-86466a6b09ef-kube-api-access-4s59t\") pod \"machine-api-operator-5694c8668f-6glz5\" (UID: \"6f28beca-7506-4390-9c74-86466a6b09ef\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.047335 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mffbs\" (UniqueName: \"kubernetes.io/projected/e9ae59fe-6c8c-4906-8c36-2b808d0474c8-kube-api-access-mffbs\") pod \"router-default-5444994796-5chwn\" (UID: \"e9ae59fe-6c8c-4906-8c36-2b808d0474c8\") " pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.076954 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2a76074e-1587-4795-a538-73d1fee5b028-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-6pxmc\" (UID: \"2a76074e-1587-4795-a538-73d1fee5b028\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.092148 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mzqp\" (UniqueName: \"kubernetes.io/projected/556bab93-68c3-4771-87cf-107ba90525ea-kube-api-access-8mzqp\") pod \"etcd-operator-b45778765-vlx6g\" (UID: \"556bab93-68c3-4771-87cf-107ba90525ea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.098371 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:33 crc kubenswrapper[4815]: E0225 13:22:33.099400 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:33.599184496 +0000 UTC m=+111.400282610 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.100472 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:33 crc kubenswrapper[4815]: E0225 13:22:33.100906 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:33.600884277 +0000 UTC m=+111.401982361 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.104017 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0281a35e-29e2-420d-97c4-dc5fb607569a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7lpx8\" (UID: \"0281a35e-29e2-420d-97c4-dc5fb607569a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lpx8" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.114480 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.125988 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.131720 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkdkn\" (UniqueName: \"kubernetes.io/projected/bf04f84d-250b-46fc-beea-9edd1c0c236a-kube-api-access-hkdkn\") pod \"openshift-apiserver-operator-796bbdcf4f-qkrlq\" (UID: \"bf04f84d-250b-46fc-beea-9edd1c0c236a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qkrlq" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.132608 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.138846 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwc6p\" (UniqueName: \"kubernetes.io/projected/5358560f-45f9-4548-8521-fad61a51f5a9-kube-api-access-vwc6p\") pod \"openshift-config-operator-7777fb866f-z7rhl\" (UID: \"5358560f-45f9-4548-8521-fad61a51f5a9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.160411 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8f2s\" (UniqueName: \"kubernetes.io/projected/5f80bb4b-ee87-4454-9335-62e45c44c5f4-kube-api-access-f8f2s\") pod \"machine-approver-56656f9798-r4kmx\" (UID: \"5f80bb4b-ee87-4454-9335-62e45c44c5f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.179827 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cd91d14c-0e3a-48a8-948f-22248711def6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qkx6p\" (UID: \"cd91d14c-0e3a-48a8-948f-22248711def6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.183799 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lpx8" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.190621 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qkrlq" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.201357 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:33 crc kubenswrapper[4815]: E0225 13:22:33.201488 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:33.701469791 +0000 UTC m=+111.502567845 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.201704 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:33 crc kubenswrapper[4815]: E0225 13:22:33.202179 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:33.702168253 +0000 UTC m=+111.503266307 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.212326 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfgsx\" (UniqueName: \"kubernetes.io/projected/e8a672b4-6113-4f9a-8d34-e60d9c8c40d2-kube-api-access-sfgsx\") pod \"apiserver-76f77b778f-hckbl\" (UID: \"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2\") " pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.221696 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z49gr\" (UniqueName: \"kubernetes.io/projected/fb4ee7c5-3dbf-4256-83e8-482b44dfea31-kube-api-access-z49gr\") pod \"migrator-59844c95c7-fhfxk\" (UID: \"fb4ee7c5-3dbf-4256-83e8-482b44dfea31\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fhfxk" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.221992 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8jnnl" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.222665 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.245246 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzf4j\" (UniqueName: \"kubernetes.io/projected/0e8fcc05-9af2-497a-8639-109eaa9e75ad-kube-api-access-kzf4j\") pod \"machine-config-operator-74547568cd-gzlvz\" (UID: \"0e8fcc05-9af2-497a-8639-109eaa9e75ad\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.260961 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsp8j\" (UniqueName: \"kubernetes.io/projected/61c629d8-0773-44ce-8919-b32d32aa5d11-kube-api-access-nsp8j\") pod \"control-plane-machine-set-operator-78cbb6b69f-x5wvt\" (UID: \"61c629d8-0773-44ce-8919-b32d32aa5d11\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5wvt" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.269370 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5wvt" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.278977 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.281396 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppmzr\" (UniqueName: \"kubernetes.io/projected/cd91d14c-0e3a-48a8-948f-22248711def6-kube-api-access-ppmzr\") pod \"ingress-operator-5b745b69d9-qkx6p\" (UID: \"cd91d14c-0e3a-48a8-948f-22248711def6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.303186 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:33 crc kubenswrapper[4815]: E0225 13:22:33.303980 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:33.80381937 +0000 UTC m=+111.604917424 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.316434 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d6t6\" (UniqueName: \"kubernetes.io/projected/60971790-c5dc-4884-ae44-e91d035cbe31-kube-api-access-2d6t6\") pod \"controller-manager-879f6c89f-4klbc\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.321844 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6a62c4db-cacf-45bb-9a8d-caefb482153b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-ff75l\" (UID: \"6a62c4db-cacf-45bb-9a8d-caefb482153b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ff75l" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.341925 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2l6w\" (UniqueName: \"kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-kube-api-access-x2l6w\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.349861 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.363076 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.373231 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-bound-sa-token\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.380168 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dv4r\" (UniqueName: \"kubernetes.io/projected/bdc0843c-6110-4d69-b444-7e5384f858e1-kube-api-access-7dv4r\") pod \"console-operator-58897d9998-w8lg6\" (UID: \"bdc0843c-6110-4d69-b444-7e5384f858e1\") " pod="openshift-console-operator/console-operator-58897d9998-w8lg6" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.386376 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.399926 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.401149 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7rzd\" (UniqueName: \"kubernetes.io/projected/71066f15-4379-475e-8295-671bfd6ea42b-kube-api-access-q7rzd\") pod \"catalog-operator-68c6474976-lttkb\" (UID: \"71066f15-4379-475e-8295-671bfd6ea42b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.404931 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:33 crc kubenswrapper[4815]: E0225 13:22:33.405246 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:33.905235359 +0000 UTC m=+111.706333413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.410745 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fhfxk" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.411571 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-w8lg6" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.422356 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs2dt\" (UniqueName: \"kubernetes.io/projected/3503ed80-0534-4d0b-a69a-fc9e9c5fa092-kube-api-access-fs2dt\") pod \"olm-operator-6b444d44fb-qdgb5\" (UID: \"3503ed80-0534-4d0b-a69a-fc9e9c5fa092\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.442442 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksc96\" (UniqueName: \"kubernetes.io/projected/7bcf8b34-940d-41d9-b32a-5548c5cc62c2-kube-api-access-ksc96\") pod \"package-server-manager-789f6589d5-njtsc\" (UID: \"7bcf8b34-940d-41d9-b32a-5548c5cc62c2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-njtsc" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.449433 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.467430 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlj7n\" (UniqueName: \"kubernetes.io/projected/ebefeb1c-aa64-4a51-8d2d-197d8e8421af-kube-api-access-mlj7n\") pod \"openshift-controller-manager-operator-756b6f6bc6-jcf7k\" (UID: \"ebefeb1c-aa64-4a51-8d2d-197d8e8421af\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jcf7k" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.487435 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brbmt\" (UniqueName: \"kubernetes.io/projected/72d966ad-816d-4483-b9e6-44edb10f9d75-kube-api-access-brbmt\") pod \"route-controller-manager-6576b87f9c-plsdd\" (UID: \"72d966ad-816d-4483-b9e6-44edb10f9d75\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.490472 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6glz5"] Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.498267 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ff75l" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.502894 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k82t\" (UniqueName: \"kubernetes.io/projected/6403af09-14ef-49fc-858f-06c1bbadb88b-kube-api-access-7k82t\") pod \"auto-csr-approver-29533762-lkzv9\" (UID: \"6403af09-14ef-49fc-858f-06c1bbadb88b\") " pod="openshift-infra/auto-csr-approver-29533762-lkzv9" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.507102 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:33 crc kubenswrapper[4815]: E0225 13:22:33.507230 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:34.007211915 +0000 UTC m=+111.808309959 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.507612 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:33 crc kubenswrapper[4815]: E0225 13:22:33.507957 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:34.007941998 +0000 UTC m=+111.809040052 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.517183 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.520630 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmrn9\" (UniqueName: \"kubernetes.io/projected/91bd28b3-4940-4027-8753-18335157566e-kube-api-access-vmrn9\") pod \"dns-operator-744455d44c-9t9rg\" (UID: \"91bd28b3-4940-4027-8753-18335157566e\") " pod="openshift-dns-operator/dns-operator-744455d44c-9t9rg" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.538642 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" event={"ID":"5f80bb4b-ee87-4454-9335-62e45c44c5f4","Type":"ContainerStarted","Data":"d1f158eca0f41ed76c35599566ae97925e9423d307682206a16a9d9c9a1762d4"} Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.538865 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lpx8"] Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.539797 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5chwn" event={"ID":"e9ae59fe-6c8c-4906-8c36-2b808d0474c8","Type":"ContainerStarted","Data":"ba2ca4f2eb03dc426dbdad37f0fcd60f198c29ae74200e408778cf5f67ba98ef"} Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.539822 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5chwn" event={"ID":"e9ae59fe-6c8c-4906-8c36-2b808d0474c8","Type":"ContainerStarted","Data":"d6d2224a40af400d5f28530d1cdb50655a9d016b250dedb32077bc1750f6700f"} Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.541207 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jcf7k" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.544848 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfh4w\" (UniqueName: \"kubernetes.io/projected/3c08c2da-42cd-47ba-92e6-aed1d48855d3-kube-api-access-lfh4w\") pod \"csi-hostpathplugin-j9qns\" (UID: \"3c08c2da-42cd-47ba-92e6-aed1d48855d3\") " pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.552430 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qkrlq"] Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.552641 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.591982 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdvqb\" (UniqueName: \"kubernetes.io/projected/770461e1-2b0c-4e47-90d1-67f2286e88e2-kube-api-access-qdvqb\") pod \"machine-config-controller-84d6567774-w2rk4\" (UID: \"770461e1-2b0c-4e47-90d1-67f2286e88e2\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.593433 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86tlk\" (UniqueName: \"kubernetes.io/projected/887d2231-1877-4977-ac7d-9f7cf1b33a23-kube-api-access-86tlk\") pod \"oauth-openshift-558db77b4-md7bn\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.604265 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.608562 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.610271 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp9x9\" (UniqueName: \"kubernetes.io/projected/601c110b-f7ba-4f1c-92ca-dc3cc6634b53-kube-api-access-zp9x9\") pod \"authentication-operator-69f744f599-tswct\" (UID: \"601c110b-f7ba-4f1c-92ca-dc3cc6634b53\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" Feb 25 13:22:33 crc kubenswrapper[4815]: E0225 13:22:33.610700 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:34.110674437 +0000 UTC m=+111.911772561 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.611649 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:33 crc kubenswrapper[4815]: E0225 13:22:33.611914 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:34.111902205 +0000 UTC m=+111.913000259 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.613283 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qscdn"] Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.615232 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.617840 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vlx6g"] Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.627445 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.636780 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc"] Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.637313 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txx4q\" (UniqueName: \"kubernetes.io/projected/6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c-kube-api-access-txx4q\") pod \"machine-config-server-ccjcb\" (UID: \"6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c\") " pod="openshift-machine-config-operator/machine-config-server-ccjcb" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.649278 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6zdr\" (UniqueName: \"kubernetes.io/projected/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-kube-api-access-c6zdr\") pod \"console-f9d7485db-mhgxd\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.669464 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsqqc\" (UniqueName: \"kubernetes.io/projected/a14e9f5d-f4ac-436b-8a6e-9523697d2b74-kube-api-access-rsqqc\") pod \"cluster-samples-operator-665b6dd947-shdc4\" (UID: \"a14e9f5d-f4ac-436b-8a6e-9523697d2b74\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-shdc4" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.672198 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-njtsc" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.680375 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggzb8\" (UniqueName: \"kubernetes.io/projected/65dfccf6-aa91-41fa-95d6-fd4778065408-kube-api-access-ggzb8\") pod \"cni-sysctl-allowlist-ds-dv7nv\" (UID: \"65dfccf6-aa91-41fa-95d6-fd4778065408\") " pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.699008 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533762-lkzv9" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.699304 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr9jx\" (UniqueName: \"kubernetes.io/projected/995e7ee9-bf81-43e8-b4d8-e11603a26e0e-kube-api-access-hr9jx\") pod \"apiserver-7bbb656c7d-qbnmz\" (UID: \"995e7ee9-bf81-43e8-b4d8-e11603a26e0e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.705718 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-shdc4" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.713158 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:33 crc kubenswrapper[4815]: E0225 13:22:33.713712 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:34.213694585 +0000 UTC m=+112.014792639 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.725826 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km5v7\" (UniqueName: \"kubernetes.io/projected/1b2fb1a6-e1d1-48e0-83dc-17143971294e-kube-api-access-km5v7\") pod \"marketplace-operator-79b997595-6th6z\" (UID: \"1b2fb1a6-e1d1-48e0-83dc-17143971294e\") " pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.726109 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-9t9rg" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.732323 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.738218 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw7jm\" (UniqueName: \"kubernetes.io/projected/08e7cd4c-1c51-4142-9384-c0d4561e7e36-kube-api-access-fw7jm\") pod \"service-ca-9c57cc56f-wdwxg\" (UID: \"08e7cd4c-1c51-4142-9384-c0d4561e7e36\") " pod="openshift-service-ca/service-ca-9c57cc56f-wdwxg" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.756132 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ccjcb" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.760956 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-j9qns" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.776219 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxkh9\" (UniqueName: \"kubernetes.io/projected/c9bef50a-04f6-427a-813c-bb0039c7aa1a-kube-api-access-zxkh9\") pod \"ingress-canary-2dz2z\" (UID: \"c9bef50a-04f6-427a-813c-bb0039c7aa1a\") " pod="openshift-ingress-canary/ingress-canary-2dz2z" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.787862 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9j6m\" (UniqueName: \"kubernetes.io/projected/d6db35c3-7baa-4452-b8ff-5c20470142d6-kube-api-access-w9j6m\") pod \"multus-admission-controller-857f4d67dd-dqj92\" (UID: \"d6db35c3-7baa-4452-b8ff-5c20470142d6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dqj92" Feb 25 13:22:33 crc kubenswrapper[4815]: W0225 13:22:33.810531 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod556bab93_68c3_4771_87cf_107ba90525ea.slice/crio-dfe813dd62a77d3e873171ba9a1a35190bdc636192f9d3cedbf0b8e6808772d4 WatchSource:0}: Error finding container dfe813dd62a77d3e873171ba9a1a35190bdc636192f9d3cedbf0b8e6808772d4: Status 404 returned error can't find the container with id dfe813dd62a77d3e873171ba9a1a35190bdc636192f9d3cedbf0b8e6808772d4 Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.815095 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r8kk\" (UniqueName: \"kubernetes.io/projected/84b7af84-2b2c-48f7-b7b0-e38fb0f7d702-kube-api-access-7r8kk\") pod \"dns-default-kw6xm\" (UID: \"84b7af84-2b2c-48f7-b7b0-e38fb0f7d702\") " pod="openshift-dns/dns-default-kw6xm" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.815236 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:33 crc kubenswrapper[4815]: E0225 13:22:33.815740 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:34.315727884 +0000 UTC m=+112.116825938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.818800 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/643cac92-ba9e-4d3c-97b7-ca525f8fc046-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-c699m\" (UID: \"643cac92-ba9e-4d3c-97b7-ca525f8fc046\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c699m" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.822991 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.831277 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8jnnl"] Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.833707 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-dqj92" Feb 25 13:22:33 crc kubenswrapper[4815]: W0225 13:22:33.837090 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a76074e_1587_4795_a538_73d1fee5b028.slice/crio-82a39ed8fbc4978a4c97c8dd93fac1f6d94219228e7d33c86ef8bbb855115e5e WatchSource:0}: Error finding container 82a39ed8fbc4978a4c97c8dd93fac1f6d94219228e7d33c86ef8bbb855115e5e: Status 404 returned error can't find the container with id 82a39ed8fbc4978a4c97c8dd93fac1f6d94219228e7d33c86ef8bbb855115e5e Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.855281 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5wvt"] Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.861179 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tslj\" (UniqueName: \"kubernetes.io/projected/739241b1-c772-4ef7-b149-e2098da43d67-kube-api-access-5tslj\") pod \"packageserver-d55dfcdfc-h7bdq\" (UID: \"739241b1-c772-4ef7-b149-e2098da43d67\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.864257 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpjhl\" (UniqueName: \"kubernetes.io/projected/87c350ba-fd2f-4e8c-a057-be035865dbf4-kube-api-access-bpjhl\") pod \"service-ca-operator-777779d784-7g5t4\" (UID: \"87c350ba-fd2f-4e8c-a057-be035865dbf4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7g5t4" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.869065 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.877773 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c699m" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.882689 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w52z\" (UniqueName: \"kubernetes.io/projected/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-kube-api-access-2w52z\") pod \"collect-profiles-29533755-65vjv\" (UID: \"80629dad-7d0a-46b3-bb52-09b75e7dfeb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.883494 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.894119 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.908493 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.919420 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:33 crc kubenswrapper[4815]: E0225 13:22:33.919871 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:34.419853496 +0000 UTC m=+112.220951560 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.926726 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.945618 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.945817 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.956518 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.964466 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wdwxg" Feb 25 13:22:33 crc kubenswrapper[4815]: I0225 13:22:33.964882 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.018983 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7g5t4" Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.022471 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:34 crc kubenswrapper[4815]: E0225 13:22:34.022877 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:34.522864494 +0000 UTC m=+112.323962548 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.040286 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-kw6xm" Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.061772 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2dz2z" Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.128975 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.129164 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:34 crc kubenswrapper[4815]: E0225 13:22:34.129239 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:34.629213583 +0000 UTC m=+112.430311637 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.129426 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:34 crc kubenswrapper[4815]: E0225 13:22:34.129946 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:34.629934785 +0000 UTC m=+112.431032839 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.141328 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl"] Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.143955 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4klbc"] Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.185352 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz"] Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.198587 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fhfxk"] Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.230265 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:34 crc kubenswrapper[4815]: E0225 13:22:34.231822 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:34.731801148 +0000 UTC m=+112.532899202 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.237071 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hckbl"] Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.333169 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:34 crc kubenswrapper[4815]: E0225 13:22:34.333663 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:34.833642851 +0000 UTC m=+112.634740915 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:34 crc kubenswrapper[4815]: W0225 13:22:34.395096 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5358560f_45f9_4548_8521_fad61a51f5a9.slice/crio-f1727baf22a1fb6d90c6f7336ac24dab853195f0dc6668c4586c207901640c10 WatchSource:0}: Error finding container f1727baf22a1fb6d90c6f7336ac24dab853195f0dc6668c4586c207901640c10: Status 404 returned error can't find the container with id f1727baf22a1fb6d90c6f7336ac24dab853195f0dc6668c4586c207901640c10 Feb 25 13:22:34 crc kubenswrapper[4815]: W0225 13:22:34.413835 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60971790_c5dc_4884_ae44_e91d035cbe31.slice/crio-b3426afcad311531ab9fc6fbdc97435f2d4bfdca83a6a49fae81620d68f1cdae WatchSource:0}: Error finding container b3426afcad311531ab9fc6fbdc97435f2d4bfdca83a6a49fae81620d68f1cdae: Status 404 returned error can't find the container with id b3426afcad311531ab9fc6fbdc97435f2d4bfdca83a6a49fae81620d68f1cdae Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.439453 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:34 crc kubenswrapper[4815]: E0225 13:22:34.440218 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:34.940196927 +0000 UTC m=+112.741294981 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.455389 4815 patch_prober.go:28] interesting pod/router-default-5444994796-5chwn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 13:22:34 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Feb 25 13:22:34 crc kubenswrapper[4815]: [+]process-running ok Feb 25 13:22:34 crc kubenswrapper[4815]: healthz check failed Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.455474 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5chwn" podUID="e9ae59fe-6c8c-4906-8c36-2b808d0474c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.541185 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:34 crc kubenswrapper[4815]: E0225 13:22:34.542486 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:35.042470712 +0000 UTC m=+112.843568766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.567554 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd"] Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.572376 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-w8lg6"] Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.586287 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p"] Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.604905 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" event={"ID":"5f80bb4b-ee87-4454-9335-62e45c44c5f4","Type":"ContainerStarted","Data":"aebb5769b1cf15a2a99538a1a2bb4c57b61febdec7aeb12e542f412e4dc4f1ea"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.604947 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" event={"ID":"5f80bb4b-ee87-4454-9335-62e45c44c5f4","Type":"ContainerStarted","Data":"eb6b45459dda07a28a2d50b8f3d22c08aa1cc6d235c838b24ca95a0790cef71b"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.610114 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc" event={"ID":"2a76074e-1587-4795-a538-73d1fee5b028","Type":"ContainerStarted","Data":"82a39ed8fbc4978a4c97c8dd93fac1f6d94219228e7d33c86ef8bbb855115e5e"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.618038 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" event={"ID":"5358560f-45f9-4548-8521-fad61a51f5a9","Type":"ContainerStarted","Data":"f1727baf22a1fb6d90c6f7336ac24dab853195f0dc6668c4586c207901640c10"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.619424 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8jnnl" event={"ID":"78d30b59-0a66-44e5-b0b9-a2864e4d4552","Type":"ContainerStarted","Data":"31862f0df02e106d06d164b351c1d0a4f940d191a324054c576c643baf739763"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.620498 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" event={"ID":"65dfccf6-aa91-41fa-95d6-fd4778065408","Type":"ContainerStarted","Data":"7fda2a9d3d2ceaabbf2867e57f172d7adbcc17c7e03f299a6fdc2377d63b8c90"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.622387 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5wvt" event={"ID":"61c629d8-0773-44ce-8919-b32d32aa5d11","Type":"ContainerStarted","Data":"4854a6998a00adb7181d4b0266c12cf83fa13fdb24b3563fc4fddc4124d76e68"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.625178 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" event={"ID":"0e8fcc05-9af2-497a-8639-109eaa9e75ad","Type":"ContainerStarted","Data":"893a676dcc8ccd2cd37c0db82bf357f8d4d5d1d89ae05310a6b1bde3271dddcc"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.626464 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qscdn" event={"ID":"a2f6b143-248f-4851-966e-93009db4876b","Type":"ContainerStarted","Data":"885c68992a1af43bf0c832098c103985d53697496146e380ce59e98df9967a3c"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.626487 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qscdn" event={"ID":"a2f6b143-248f-4851-966e-93009db4876b","Type":"ContainerStarted","Data":"ab99b561c6e5c754f27756aace6e531f481256ad3b0e44436b147e70641377bd"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.627830 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fhfxk" event={"ID":"fb4ee7c5-3dbf-4256-83e8-482b44dfea31","Type":"ContainerStarted","Data":"48402e18dd6d5dff33802744b110291e462413be73f3245158921cdd71bd2768"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.628947 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" event={"ID":"60971790-c5dc-4884-ae44-e91d035cbe31","Type":"ContainerStarted","Data":"b3426afcad311531ab9fc6fbdc97435f2d4bfdca83a6a49fae81620d68f1cdae"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.631331 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" event={"ID":"6f28beca-7506-4390-9c74-86466a6b09ef","Type":"ContainerStarted","Data":"492e06cc2f1169727f8346d6d1c801c36377203fadb878be3111539dd5bdf23f"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.631355 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" event={"ID":"6f28beca-7506-4390-9c74-86466a6b09ef","Type":"ContainerStarted","Data":"eb48a60f96364b19d1f3ce6716dfe1f3fd71cf7bc424775fb5460ba1e7fab1f7"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.632310 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" event={"ID":"556bab93-68c3-4771-87cf-107ba90525ea","Type":"ContainerStarted","Data":"dfe813dd62a77d3e873171ba9a1a35190bdc636192f9d3cedbf0b8e6808772d4"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.633613 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ccjcb" event={"ID":"6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c","Type":"ContainerStarted","Data":"14c63490a07737d2ff3c98331839fb43d14402049c7b158f4bb1fff6a7a85136"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.640231 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hckbl" event={"ID":"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2","Type":"ContainerStarted","Data":"09eee474b1d7e14810e0da5c7cb3afd529779e8bbdb34f2f71e01caedbc9d340"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.641841 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:34 crc kubenswrapper[4815]: E0225 13:22:34.641943 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:35.141919742 +0000 UTC m=+112.943017806 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.642198 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:34 crc kubenswrapper[4815]: E0225 13:22:34.642478 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:35.142467049 +0000 UTC m=+112.943565103 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.643335 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lpx8" event={"ID":"0281a35e-29e2-420d-97c4-dc5fb607569a","Type":"ContainerStarted","Data":"42fa9fc859a36c983b9fef5144f7030235e5a7ad8cc596fd01c7409516d332b8"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.643357 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lpx8" event={"ID":"0281a35e-29e2-420d-97c4-dc5fb607569a","Type":"ContainerStarted","Data":"6b08a8d873411c5c7e4bc241794c189847baa8daae271f465f41ec1c1a5fd596"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.646286 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qkrlq" event={"ID":"bf04f84d-250b-46fc-beea-9edd1c0c236a","Type":"ContainerStarted","Data":"ba683510bf8236d7813baf03df55d13200bc74589fbd620fd1a78b97ae957745"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.646355 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qkrlq" event={"ID":"bf04f84d-250b-46fc-beea-9edd1c0c236a","Type":"ContainerStarted","Data":"aa68f42c9bd5c97bee4607606e264d84110774146866149e632482f420aa9d83"} Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.743225 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:34 crc kubenswrapper[4815]: E0225 13:22:34.745151 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:35.245112176 +0000 UTC m=+113.046210240 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.771884 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jcf7k"] Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.852561 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:34 crc kubenswrapper[4815]: E0225 13:22:34.852935 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:35.352918209 +0000 UTC m=+113.154016333 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.936921 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb"] Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.959737 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:34 crc kubenswrapper[4815]: E0225 13:22:34.960277 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:35.460258739 +0000 UTC m=+113.261356793 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.976468 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4"] Feb 25 13:22:34 crc kubenswrapper[4815]: I0225 13:22:34.988886 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ff75l"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.061837 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:35 crc kubenswrapper[4815]: E0225 13:22:35.071017 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:35.570961331 +0000 UTC m=+113.372059385 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.118893 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-5chwn" podStartSLOduration=60.118870765 podStartE2EDuration="1m0.118870765s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:35.118828693 +0000 UTC m=+112.919926747" watchObservedRunningTime="2026-02-25 13:22:35.118870765 +0000 UTC m=+112.919968829" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.135021 4815 patch_prober.go:28] interesting pod/router-default-5444994796-5chwn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 13:22:35 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Feb 25 13:22:35 crc kubenswrapper[4815]: [+]process-running ok Feb 25 13:22:35 crc kubenswrapper[4815]: healthz check failed Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.135062 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5chwn" podUID="e9ae59fe-6c8c-4906-8c36-2b808d0474c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.161178 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-md7bn"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.163114 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:35 crc kubenswrapper[4815]: E0225 13:22:35.163819 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:35.663799459 +0000 UTC m=+113.464897523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.207686 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.210688 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dqj92"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.213837 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-shdc4"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.214496 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mhgxd"] Feb 25 13:22:35 crc kubenswrapper[4815]: W0225 13:22:35.242641 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6db35c3_7baa_4452_b8ff_5c20470142d6.slice/crio-f587a541b0557bf2f9dd1bb40cf312aea6a69097eadb85d58cf38894d3dd3ec8 WatchSource:0}: Error finding container f587a541b0557bf2f9dd1bb40cf312aea6a69097eadb85d58cf38894d3dd3ec8: Status 404 returned error can't find the container with id f587a541b0557bf2f9dd1bb40cf312aea6a69097eadb85d58cf38894d3dd3ec8 Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.253396 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-njtsc"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.253440 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9t9rg"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.266847 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:35 crc kubenswrapper[4815]: E0225 13:22:35.267214 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:35.767198829 +0000 UTC m=+113.568296883 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.289359 4815 ???:1] "http: TLS handshake error from 192.168.126.11:50490: no serving certificate available for the kubelet" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.289865 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-j9qns"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.295054 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c699m"] Feb 25 13:22:35 crc kubenswrapper[4815]: W0225 13:22:35.304661 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91bd28b3_4940_4027_8753_18335157566e.slice/crio-88b8acb02e2d52f7a41e71e87fa895ef242bfe93e55a2c06158f62d9703c2756 WatchSource:0}: Error finding container 88b8acb02e2d52f7a41e71e87fa895ef242bfe93e55a2c06158f62d9703c2756: Status 404 returned error can't find the container with id 88b8acb02e2d52f7a41e71e87fa895ef242bfe93e55a2c06158f62d9703c2756 Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.305098 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.317487 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6th6z"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.334122 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533762-lkzv9"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.339759 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wdwxg"] Feb 25 13:22:35 crc kubenswrapper[4815]: W0225 13:22:35.343714 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c08c2da_42cd_47ba_92e6_aed1d48855d3.slice/crio-48144638ecd21e68debe2d23a0cc64b7f4769fa6fff3f261390cff631cc9f05d WatchSource:0}: Error finding container 48144638ecd21e68debe2d23a0cc64b7f4769fa6fff3f261390cff631cc9f05d: Status 404 returned error can't find the container with id 48144638ecd21e68debe2d23a0cc64b7f4769fa6fff3f261390cff631cc9f05d Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.349818 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.354326 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.355632 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-kw6xm"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.362478 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2dz2z"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.368487 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:35 crc kubenswrapper[4815]: E0225 13:22:35.368917 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:35.868903268 +0000 UTC m=+113.670001322 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.370110 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7g5t4"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.375743 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-tswct"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.405059 4815 ???:1] "http: TLS handshake error from 192.168.126.11:50506: no serving certificate available for the kubelet" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.435543 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.482999 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:35 crc kubenswrapper[4815]: E0225 13:22:35.483279 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:35.98326868 +0000 UTC m=+113.784366734 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.495405 4815 ???:1] "http: TLS handshake error from 192.168.126.11:50520: no serving certificate available for the kubelet" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.584577 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:35 crc kubenswrapper[4815]: E0225 13:22:35.584775 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:36.084754452 +0000 UTC m=+113.885852496 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.585039 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:35 crc kubenswrapper[4815]: E0225 13:22:35.585609 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:36.085575597 +0000 UTC m=+113.886673651 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.609677 4815 ???:1] "http: TLS handshake error from 192.168.126.11:50524: no serving certificate available for the kubelet" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.658938 4815 generic.go:334] "Generic (PLEG): container finished" podID="e8a672b4-6113-4f9a-8d34-e60d9c8c40d2" containerID="4d31d313385c7f3af29d155f041e07d0d879600c2c21852658ed83199b4f5c8a" exitCode=0 Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.659038 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hckbl" event={"ID":"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2","Type":"ContainerDied","Data":"4d31d313385c7f3af29d155f041e07d0d879600c2c21852658ed83199b4f5c8a"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.661331 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" event={"ID":"1b2fb1a6-e1d1-48e0-83dc-17143971294e","Type":"ContainerStarted","Data":"f447cce8a81086097513032f03607144a55ef1b9b23ec49f2a727d8a96b55a04"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.666214 4815 generic.go:334] "Generic (PLEG): container finished" podID="5358560f-45f9-4548-8521-fad61a51f5a9" containerID="b3600bbf4ab3bd5d94a17406c6055369a68765a4786e0c6aa28a81d3602fc9c2" exitCode=0 Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.666326 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" event={"ID":"5358560f-45f9-4548-8521-fad61a51f5a9","Type":"ContainerDied","Data":"b3600bbf4ab3bd5d94a17406c6055369a68765a4786e0c6aa28a81d3602fc9c2"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.674069 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jcf7k" event={"ID":"ebefeb1c-aa64-4a51-8d2d-197d8e8421af","Type":"ContainerStarted","Data":"ca6b8b32cc9039110cc6fbb50c84d724abc434f8de547e4843d9906da3bc3585"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.674110 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jcf7k" event={"ID":"ebefeb1c-aa64-4a51-8d2d-197d8e8421af","Type":"ContainerStarted","Data":"bdb59d54c31ac592041fc21c3d257ac484916e9a5f7b4c5d6c5dc71a17930d9c"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.685845 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.685973 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs\") pod \"network-metrics-daemon-gbqbs\" (UID: \"97b044d0-af42-4f42-95a8-49d6c8782246\") " pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:35 crc kubenswrapper[4815]: E0225 13:22:35.686043 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:36.185998987 +0000 UTC m=+113.987097091 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.690684 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ccjcb" event={"ID":"6e0b1cf5-40e4-476c-8e9f-0e81e3ce860c","Type":"ContainerStarted","Data":"e14bc29d627f7ef4cd884a6148a3cbc06641c91af79b97e57b9075b9f80c0ac2"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.693484 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ff75l" event={"ID":"6a62c4db-cacf-45bb-9a8d-caefb482153b","Type":"ContainerStarted","Data":"2b6d73d63318a17194142c913d6173448c5d1e4bd5dc0f481aeaef14dc62f5ec"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.697182 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" event={"ID":"601c110b-f7ba-4f1c-92ca-dc3cc6634b53","Type":"ContainerStarted","Data":"c5f7085cf8c9d6d1ef8cad42d93227cffe33a5dfeedd8f1de2126cc81175338e"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.701108 4815 ???:1] "http: TLS handshake error from 192.168.126.11:50526: no serving certificate available for the kubelet" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.716211 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/97b044d0-af42-4f42-95a8-49d6c8782246-metrics-certs\") pod \"network-metrics-daemon-gbqbs\" (UID: \"97b044d0-af42-4f42-95a8-49d6c8782246\") " pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.722534 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qscdn" podStartSLOduration=60.722516216 podStartE2EDuration="1m0.722516216s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:35.721247717 +0000 UTC m=+113.522345771" watchObservedRunningTime="2026-02-25 13:22:35.722516216 +0000 UTC m=+113.523614270" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.726536 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5" event={"ID":"3503ed80-0534-4d0b-a69a-fc9e9c5fa092","Type":"ContainerStarted","Data":"a1d968963d0e90fd4a68b5676db5af879d165b2280000555c1714d163c8a403e"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.726587 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5" event={"ID":"3503ed80-0534-4d0b-a69a-fc9e9c5fa092","Type":"ContainerStarted","Data":"2b94cd3c60a451cf3903a7470cfef80af759793feed0e680354136fd163f9d75"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.727476 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.732526 4815 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-qdgb5 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.732570 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5" podUID="3503ed80-0534-4d0b-a69a-fc9e9c5fa092" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.748747 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gbqbs" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.756157 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" event={"ID":"0e8fcc05-9af2-497a-8639-109eaa9e75ad","Type":"ContainerStarted","Data":"e1db9595afeb1e36f104f929586ab1b99e96921556f1222e824f619c9137baa1"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.756217 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" event={"ID":"0e8fcc05-9af2-497a-8639-109eaa9e75ad","Type":"ContainerStarted","Data":"00676e68a5649a689c8f8df1f811b1248f6bf9525876c6436d7777587314c4e8"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.768190 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-j9qns" event={"ID":"3c08c2da-42cd-47ba-92e6-aed1d48855d3","Type":"ContainerStarted","Data":"48144638ecd21e68debe2d23a0cc64b7f4769fa6fff3f261390cff631cc9f05d"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.777860 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8jnnl" event={"ID":"78d30b59-0a66-44e5-b0b9-a2864e4d4552","Type":"ContainerStarted","Data":"3e9dee7d7e1a9d47ddda8837fc98161fe623d7a55f8f7c164ef85d9a3c10c1c8"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.778936 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-8jnnl" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.782014 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-8jnnl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" start-of-body= Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.782065 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8jnnl" podUID="78d30b59-0a66-44e5-b0b9-a2864e4d4552" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.789282 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:35 crc kubenswrapper[4815]: E0225 13:22:35.790703 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:36.290688165 +0000 UTC m=+114.091786219 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.791211 4815 ???:1] "http: TLS handshake error from 192.168.126.11:50532: no serving certificate available for the kubelet" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.794251 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kw6xm" event={"ID":"84b7af84-2b2c-48f7-b7b0-e38fb0f7d702","Type":"ContainerStarted","Data":"7c76578b7770eae34ab383d1c0e994ed184b70709c0452ad15361784dd9b20b6"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.802664 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" event={"ID":"6f28beca-7506-4390-9c74-86466a6b09ef","Type":"ContainerStarted","Data":"f155f7fa7fecb2e61ffe125adb2a172e528eda2fdbdd25a423752e80dd925247"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.808420 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" event={"ID":"80629dad-7d0a-46b3-bb52-09b75e7dfeb9","Type":"ContainerStarted","Data":"28a7fa6c0beba75fc11d190790b4525724e33b78a6d798bacdab35592dc0971c"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.814658 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mhgxd" event={"ID":"a81edad3-179a-4786-9ae9-e1a8f0a4d01a","Type":"ContainerStarted","Data":"0089f754c2c276e7a440758d91f196d27023f0649f2625a873528e9077ce8a42"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.824154 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lpx8" podStartSLOduration=60.824134652 podStartE2EDuration="1m0.824134652s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:35.823056519 +0000 UTC m=+113.624154573" watchObservedRunningTime="2026-02-25 13:22:35.824134652 +0000 UTC m=+113.625232706" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.827125 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-njtsc" event={"ID":"7bcf8b34-940d-41d9-b32a-5548c5cc62c2","Type":"ContainerStarted","Data":"97df392210c5b6785004152e9ec281b9490e556a2db56a23b444f16140aeb271"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.841278 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-shdc4" event={"ID":"a14e9f5d-f4ac-436b-8a6e-9523697d2b74","Type":"ContainerStarted","Data":"f9d205d68457e18156de72d74f1e1f4a816fc76aa978a474ca990bdfada3bc2e"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.850448 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4" event={"ID":"770461e1-2b0c-4e47-90d1-67f2286e88e2","Type":"ContainerStarted","Data":"ec3000f22c342e3dda520fa188d0a73699bef9a6824bdd3cb0ba151323672912"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.850725 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4" event={"ID":"770461e1-2b0c-4e47-90d1-67f2286e88e2","Type":"ContainerStarted","Data":"59dd61d031c0eb5f884b8b4a7957af3081ec9eb3b57d2d975d0d4e8adc0c1fcd"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.852583 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" event={"ID":"995e7ee9-bf81-43e8-b4d8-e11603a26e0e","Type":"ContainerStarted","Data":"34a17d7c35dcb755600d929a110b9f1a023dc34cefe99dcd291251acd0256fdf"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.857873 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb" event={"ID":"71066f15-4379-475e-8295-671bfd6ea42b","Type":"ContainerStarted","Data":"8d16232854cf728c3debc0d3ff6552f44b26de426b9acc54162d4e0f50d87229"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.857915 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb" event={"ID":"71066f15-4379-475e-8295-671bfd6ea42b","Type":"ContainerStarted","Data":"79351b0e358a24a388330acbd9f8f338dc8aeea989a026100cd618f21c99a3aa"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.858549 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.859598 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9t9rg" event={"ID":"91bd28b3-4940-4027-8753-18335157566e","Type":"ContainerStarted","Data":"88b8acb02e2d52f7a41e71e87fa895ef242bfe93e55a2c06158f62d9703c2756"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.861937 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c699m" event={"ID":"643cac92-ba9e-4d3c-97b7-ca525f8fc046","Type":"ContainerStarted","Data":"d9c0836d6ae122d71d0b1db77684578bda496e43a78444c0f57515cf01584aeb"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.867373 4815 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-lttkb container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.867455 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb" podUID="71066f15-4379-475e-8295-671bfd6ea42b" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.870005 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fhfxk" event={"ID":"fb4ee7c5-3dbf-4256-83e8-482b44dfea31","Type":"ContainerStarted","Data":"0748f156e1a9631a24942f12c1833b87c2554f18e9a758bdd30a930dc6f1f41b"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.870063 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fhfxk" event={"ID":"fb4ee7c5-3dbf-4256-83e8-482b44dfea31","Type":"ContainerStarted","Data":"8c138a0557104f584de7b4cf37c5e3a3fbd672eea170337e7eda68ab05eb0d29"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.874686 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" event={"ID":"739241b1-c772-4ef7-b149-e2098da43d67","Type":"ContainerStarted","Data":"2f7b3417b24ef7de17bf11e0f8ebd07db308c83381fbf76c1b4a5af55025170c"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.880893 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wdwxg" event={"ID":"08e7cd4c-1c51-4142-9384-c0d4561e7e36","Type":"ContainerStarted","Data":"841bed063757a6dcd6e37705ab6f6a334e42ba33aa84ef875836785b02b1c460"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.884101 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" event={"ID":"887d2231-1877-4977-ac7d-9f7cf1b33a23","Type":"ContainerStarted","Data":"e9e01707c67d715db60c7a341dfa3799191bef99c3a9bc52eb77adaa3b2f4e67"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.891055 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.893806 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" event={"ID":"65dfccf6-aa91-41fa-95d6-fd4778065408","Type":"ContainerStarted","Data":"bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.894475 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:22:35 crc kubenswrapper[4815]: E0225 13:22:35.894592 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:36.39457253 +0000 UTC m=+114.195670684 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.899807 4815 ???:1] "http: TLS handshake error from 192.168.126.11:50536: no serving certificate available for the kubelet" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.901821 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5wvt" event={"ID":"61c629d8-0773-44ce-8919-b32d32aa5d11","Type":"ContainerStarted","Data":"98bcf58ba5f95dcea1e1707e4a9dac00c5c108ff3b2d3aa2da250f539ca54a4a"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.915384 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" event={"ID":"cd91d14c-0e3a-48a8-948f-22248711def6","Type":"ContainerStarted","Data":"eab531ac825ac25bf6500b345b9ee7a2038470f164518485246cb6a9a0ef43fc"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.915442 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" event={"ID":"cd91d14c-0e3a-48a8-948f-22248711def6","Type":"ContainerStarted","Data":"24b8db4ad893ace4c736f534a401e8de4d85dd62f13b8f9025b3139ef9c81099"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.923231 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dqj92" event={"ID":"d6db35c3-7baa-4452-b8ff-5c20470142d6","Type":"ContainerStarted","Data":"f587a541b0557bf2f9dd1bb40cf312aea6a69097eadb85d58cf38894d3dd3ec8"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.928961 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" event={"ID":"72d966ad-816d-4483-b9e6-44edb10f9d75","Type":"ContainerStarted","Data":"91dfc815ac31db022e98712effb6e15a11dfc06a27f450d84b1c5fa9c3d9e40a"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.929045 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" event={"ID":"72d966ad-816d-4483-b9e6-44edb10f9d75","Type":"ContainerStarted","Data":"08c83bfc91ff7986916e4fdc443458a7467f3928ad09a3629eb141adebe663f2"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.929580 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.931017 4815 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-plsdd container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.931065 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" podUID="72d966ad-816d-4483-b9e6-44edb10f9d75" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.933781 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2dz2z" event={"ID":"c9bef50a-04f6-427a-813c-bb0039c7aa1a","Type":"ContainerStarted","Data":"6e1c1c2bc406bc0b70e4c33355da48dcdd3cf7f7967d4c585ac196c985cd10b9"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.939527 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" event={"ID":"60971790-c5dc-4884-ae44-e91d035cbe31","Type":"ContainerStarted","Data":"8bd4854abbc245cd236b0340b06154edf75ebec05eb8c528b58affe3c748103b"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.939743 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.941425 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7g5t4" event={"ID":"87c350ba-fd2f-4e8c-a057-be035865dbf4","Type":"ContainerStarted","Data":"93c3db50c983ee6ed4da854d4a0ea51525f26795e247fd8cd9966d8f929c48b0"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.941784 4815 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-4klbc container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.941816 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" podUID="60971790-c5dc-4884-ae44-e91d035cbe31" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.944119 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" event={"ID":"556bab93-68c3-4771-87cf-107ba90525ea","Type":"ContainerStarted","Data":"3eac7b106f20f185de8f027982c4358edd9cf8006f1ba73d7bb824b5816479e3"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.961262 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.964132 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-w8lg6" event={"ID":"bdc0843c-6110-4d69-b444-7e5384f858e1","Type":"ContainerStarted","Data":"79d10d9d01556e9e77b042c3ba762114b70561e01457b20ebc05c5633789a99a"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.964173 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-w8lg6" event={"ID":"bdc0843c-6110-4d69-b444-7e5384f858e1","Type":"ContainerStarted","Data":"f0accf6228101104223b4bddd6118bfe410104531d3e2e6802ff1f006ed18e89"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.965798 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-w8lg6" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.970717 4815 patch_prober.go:28] interesting pod/console-operator-58897d9998-w8lg6 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.970722 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533762-lkzv9" event={"ID":"6403af09-14ef-49fc-858f-06c1bbadb88b","Type":"ContainerStarted","Data":"25a59f5eddb13f98136269efa8e1897b0426c09398f08d2b9389d2e774e53839"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.970836 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-w8lg6" podUID="bdc0843c-6110-4d69-b444-7e5384f858e1" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.979188 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc" event={"ID":"2a76074e-1587-4795-a538-73d1fee5b028","Type":"ContainerStarted","Data":"2868d6598a400221afc15209bdf02ed4a6fec290f87d818145b1dd99491e8a37"} Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.998600 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:35 crc kubenswrapper[4815]: I0225 13:22:35.999649 4815 ???:1] "http: TLS handshake error from 192.168.126.11:50538: no serving certificate available for the kubelet" Feb 25 13:22:36 crc kubenswrapper[4815]: E0225 13:22:36.000397 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:36.500384614 +0000 UTC m=+114.301482678 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.100185 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:36 crc kubenswrapper[4815]: E0225 13:22:36.101578 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:36.601563525 +0000 UTC m=+114.402661579 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.137655 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.147455 4815 patch_prober.go:28] interesting pod/router-default-5444994796-5chwn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 13:22:36 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Feb 25 13:22:36 crc kubenswrapper[4815]: [+]process-running ok Feb 25 13:22:36 crc kubenswrapper[4815]: healthz check failed Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.147524 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5chwn" podUID="e9ae59fe-6c8c-4906-8c36-2b808d0474c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.172471 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jcf7k" podStartSLOduration=61.172451388 podStartE2EDuration="1m1.172451388s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:36.139819448 +0000 UTC m=+113.940917522" watchObservedRunningTime="2026-02-25 13:22:36.172451388 +0000 UTC m=+113.973549442" Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.202291 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:36 crc kubenswrapper[4815]: E0225 13:22:36.202658 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:36.702642815 +0000 UTC m=+114.503740869 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.308035 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:36 crc kubenswrapper[4815]: E0225 13:22:36.308617 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:36.808603032 +0000 UTC m=+114.609701086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.410359 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:36 crc kubenswrapper[4815]: E0225 13:22:36.410734 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:36.910718853 +0000 UTC m=+114.711816907 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.512952 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:36 crc kubenswrapper[4815]: E0225 13:22:36.513284 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:37.013270597 +0000 UTC m=+114.814368651 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.592164 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-r4kmx" podStartSLOduration=61.592146232 podStartE2EDuration="1m1.592146232s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:36.531960395 +0000 UTC m=+114.333058459" watchObservedRunningTime="2026-02-25 13:22:36.592146232 +0000 UTC m=+114.393244286" Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.597273 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gbqbs"] Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.614683 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:36 crc kubenswrapper[4815]: E0225 13:22:36.615319 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:37.115306916 +0000 UTC m=+114.916404970 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.660450 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qkrlq" podStartSLOduration=61.660430376 podStartE2EDuration="1m1.660430376s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:36.65563554 +0000 UTC m=+114.456733594" watchObservedRunningTime="2026-02-25 13:22:36.660430376 +0000 UTC m=+114.461528430" Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.706673 4815 ???:1] "http: TLS handshake error from 192.168.126.11:50554: no serving certificate available for the kubelet" Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.723210 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:36 crc kubenswrapper[4815]: E0225 13:22:36.723468 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:37.22345461 +0000 UTC m=+115.024552654 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.733430 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" podStartSLOduration=60.733406862 podStartE2EDuration="1m0.733406862s" podCreationTimestamp="2026-02-25 13:21:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:36.713188408 +0000 UTC m=+114.514286462" watchObservedRunningTime="2026-02-25 13:22:36.733406862 +0000 UTC m=+114.534504916" Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.758498 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" podStartSLOduration=61.758476744 podStartE2EDuration="1m1.758476744s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:36.755011228 +0000 UTC m=+114.556109282" watchObservedRunningTime="2026-02-25 13:22:36.758476744 +0000 UTC m=+114.559574798" Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.775716 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6pxmc" podStartSLOduration=61.775699197 podStartE2EDuration="1m1.775699197s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:36.772599542 +0000 UTC m=+114.573697596" watchObservedRunningTime="2026-02-25 13:22:36.775699197 +0000 UTC m=+114.576797251" Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.812949 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" podStartSLOduration=6.8129296870000005 podStartE2EDuration="6.812929687s" podCreationTimestamp="2026-02-25 13:22:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:36.805333886 +0000 UTC m=+114.606431940" watchObservedRunningTime="2026-02-25 13:22:36.812929687 +0000 UTC m=+114.614027741" Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.824675 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:36 crc kubenswrapper[4815]: E0225 13:22:36.825112 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:37.325096206 +0000 UTC m=+115.126194260 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.847692 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-vlx6g" podStartSLOduration=61.847680582 podStartE2EDuration="1m1.847680582s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:36.84595385 +0000 UTC m=+114.647051904" watchObservedRunningTime="2026-02-25 13:22:36.847680582 +0000 UTC m=+114.648778636" Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.925322 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:36 crc kubenswrapper[4815]: E0225 13:22:36.925578 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:37.425550287 +0000 UTC m=+115.226648341 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.925815 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:36 crc kubenswrapper[4815]: E0225 13:22:36.926108 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:37.426100524 +0000 UTC m=+115.227198578 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.932970 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-ccjcb" podStartSLOduration=6.932955541 podStartE2EDuration="6.932955541s" podCreationTimestamp="2026-02-25 13:22:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:36.882435107 +0000 UTC m=+114.683533161" watchObservedRunningTime="2026-02-25 13:22:36.932955541 +0000 UTC m=+114.734053585" Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.972740 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=1.972726569 podStartE2EDuration="1.972726569s" podCreationTimestamp="2026-02-25 13:22:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:36.971330707 +0000 UTC m=+114.772428761" watchObservedRunningTime="2026-02-25 13:22:36.972726569 +0000 UTC m=+114.773824623" Feb 25 13:22:36 crc kubenswrapper[4815]: I0225 13:22:36.973889 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fhfxk" podStartSLOduration=61.973884785 podStartE2EDuration="1m1.973884785s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:36.943587035 +0000 UTC m=+114.744685089" watchObservedRunningTime="2026-02-25 13:22:36.973884785 +0000 UTC m=+114.774982829" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.014686 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" event={"ID":"887d2231-1877-4977-ac7d-9f7cf1b33a23","Type":"ContainerStarted","Data":"9ef793a39cc3472a5cd88bcb5c1ee71ede84f400225e179b7567a195dd768bf3"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.015352 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.017724 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4" event={"ID":"770461e1-2b0c-4e47-90d1-67f2286e88e2","Type":"ContainerStarted","Data":"0416811a942683d213816d949cdcb4e8836cb6ffe3f3586d4c3433eee9634782"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.024937 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gbqbs" event={"ID":"97b044d0-af42-4f42-95a8-49d6c8782246","Type":"ContainerStarted","Data":"b763f221517dd5e4dcb2e8ab4a13349e2cdd2c6b91a8c001a9082b4d25c324a3"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.027689 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:37 crc kubenswrapper[4815]: E0225 13:22:37.028350 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:37.528334597 +0000 UTC m=+115.329432651 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.032786 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ff75l" event={"ID":"6a62c4db-cacf-45bb-9a8d-caefb482153b","Type":"ContainerStarted","Data":"f290f93932c05c8511ea85e092bdb39e120d30155f487fdf9c93e2611ebe2f6e"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.034259 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-gzlvz" podStartSLOduration=62.034206546 podStartE2EDuration="1m2.034206546s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.025554953 +0000 UTC m=+114.826653017" watchObservedRunningTime="2026-02-25 13:22:37.034206546 +0000 UTC m=+114.835304620" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.038143 4815 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-md7bn container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.37:6443/healthz\": dial tcp 10.217.0.37:6443: connect: connection refused" start-of-body= Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.038203 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" podUID="887d2231-1877-4977-ac7d-9f7cf1b33a23" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.37:6443/healthz\": dial tcp 10.217.0.37:6443: connect: connection refused" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.064202 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-shdc4" event={"ID":"a14e9f5d-f4ac-436b-8a6e-9523697d2b74","Type":"ContainerStarted","Data":"9794f62cd3162409acec2e570a687ae926a69968dad9754c95aae70c16c681a4"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.078726 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7g5t4" event={"ID":"87c350ba-fd2f-4e8c-a057-be035865dbf4","Type":"ContainerStarted","Data":"72da2c58806c3e336d3b58ab3b03f29f2b6d96e7500a4f6518613b06cf970c06"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.092643 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" podStartSLOduration=62.0926238 podStartE2EDuration="1m2.0926238s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.091148645 +0000 UTC m=+114.892246729" watchObservedRunningTime="2026-02-25 13:22:37.0926238 +0000 UTC m=+114.893721854" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.095453 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9t9rg" event={"ID":"91bd28b3-4940-4027-8753-18335157566e","Type":"ContainerStarted","Data":"db20b2a3148f1155ae644fcab378eb22a07861d2bd1f00f0a3e2ce2652cf263f"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.095537 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9t9rg" event={"ID":"91bd28b3-4940-4027-8753-18335157566e","Type":"ContainerStarted","Data":"73b804d5f310269eac25ddab8a2bfb8df5ca234e176bb570948161e71aad8b78"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.113376 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" event={"ID":"5358560f-45f9-4548-8521-fad61a51f5a9","Type":"ContainerStarted","Data":"0c606fb4bddf022813bc7fc3e6ed882dee922d4b180e84ead7526eab1fe6357f"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.113520 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.128928 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:37 crc kubenswrapper[4815]: E0225 13:22:37.130579 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:37.630559552 +0000 UTC m=+115.431657686 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.133721 4815 patch_prober.go:28] interesting pod/router-default-5444994796-5chwn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 13:22:37 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Feb 25 13:22:37 crc kubenswrapper[4815]: [+]process-running ok Feb 25 13:22:37 crc kubenswrapper[4815]: healthz check failed Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.133762 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5chwn" podUID="e9ae59fe-6c8c-4906-8c36-2b808d0474c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.137794 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2dz2z" event={"ID":"c9bef50a-04f6-427a-813c-bb0039c7aa1a","Type":"ContainerStarted","Data":"bf3de5fe49e04c3c08f7f264363cb2d613a215ad96719462cf1f0da50a5c529f"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.139824 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-w8lg6" podStartSLOduration=62.139806473 podStartE2EDuration="1m2.139806473s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.124599091 +0000 UTC m=+114.925697145" watchObservedRunningTime="2026-02-25 13:22:37.139806473 +0000 UTC m=+114.940904527" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.149978 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-mhgxd" podStartSLOduration=62.149962341 podStartE2EDuration="1m2.149962341s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.147898679 +0000 UTC m=+114.948996733" watchObservedRunningTime="2026-02-25 13:22:37.149962341 +0000 UTC m=+114.951060395" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.174977 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5" podStartSLOduration=62.17495909 podStartE2EDuration="1m2.17495909s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.170463704 +0000 UTC m=+114.971561768" watchObservedRunningTime="2026-02-25 13:22:37.17495909 +0000 UTC m=+114.976057144" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.208099 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" event={"ID":"601c110b-f7ba-4f1c-92ca-dc3cc6634b53","Type":"ContainerStarted","Data":"7214f83fc4239d75c3cd6eed3a7aa84ca598b49190472bab5ec10648df64956a"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.234160 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:37 crc kubenswrapper[4815]: E0225 13:22:37.301551 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:37.801525974 +0000 UTC m=+115.602624028 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.518450 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.525046 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-8jnnl" podStartSLOduration=62.525029766 podStartE2EDuration="1m2.525029766s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.204895829 +0000 UTC m=+115.005993903" watchObservedRunningTime="2026-02-25 13:22:37.525029766 +0000 UTC m=+115.326127820" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.526441 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qkx6p" event={"ID":"cd91d14c-0e3a-48a8-948f-22248711def6","Type":"ContainerStarted","Data":"f17ab6ae5f149e8e01c78d18981fbbc8c5ec270246583276f9c36969031c77ac"} Feb 25 13:22:37 crc kubenswrapper[4815]: E0225 13:22:37.528450 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:38.028430389 +0000 UTC m=+115.829528523 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.578462 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb" podStartSLOduration=62.578442823 podStartE2EDuration="1m2.578442823s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.576382541 +0000 UTC m=+115.377480605" watchObservedRunningTime="2026-02-25 13:22:37.578442823 +0000 UTC m=+115.379540867" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.581645 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5wvt" podStartSLOduration=62.58162848 podStartE2EDuration="1m2.58162848s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.527187131 +0000 UTC m=+115.328285205" watchObservedRunningTime="2026-02-25 13:22:37.58162848 +0000 UTC m=+115.382726544" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.590994 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hckbl" event={"ID":"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2","Type":"ContainerStarted","Data":"ed1fc853c3ca4ff7712debe99724eecb273996296721c47cf85e29a2c0d8cf44"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.608706 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" event={"ID":"80629dad-7d0a-46b3-bb52-09b75e7dfeb9","Type":"ContainerStarted","Data":"5fcc6127de7c4f26b8ab72ffdab14f84fee434383c588d93cc0b01819e27c9cb"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.609788 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-6glz5" podStartSLOduration=62.609772564 podStartE2EDuration="1m2.609772564s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.608775074 +0000 UTC m=+115.409873128" watchObservedRunningTime="2026-02-25 13:22:37.609772564 +0000 UTC m=+115.410870618" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.614246 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mhgxd" event={"ID":"a81edad3-179a-4786-9ae9-e1a8f0a4d01a","Type":"ContainerStarted","Data":"42fd30cefd32dca44fba88b4d91e294bcfe2c963e00cd0a323bb898e8e032242"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.616704 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-njtsc" event={"ID":"7bcf8b34-940d-41d9-b32a-5548c5cc62c2","Type":"ContainerStarted","Data":"04c0c66a271fd17d0edc1c212cf6a153f16411197ef1f1c76a9f1f5a276b2034"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.616730 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-njtsc" event={"ID":"7bcf8b34-940d-41d9-b32a-5548c5cc62c2","Type":"ContainerStarted","Data":"deb109d12f240d3f7a7932a7260e1d6c18751afacfe43de1432dca78cbe4da66"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.617217 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-njtsc" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.619165 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:37 crc kubenswrapper[4815]: E0225 13:22:37.620422 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:38.120398957 +0000 UTC m=+115.921497011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.624063 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wdwxg" event={"ID":"08e7cd4c-1c51-4142-9384-c0d4561e7e36","Type":"ContainerStarted","Data":"c91b20d030a3da82c0f74a61dbe010d9729edf6a7e73b946ee3d10b9746eca29"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.629951 4815 generic.go:334] "Generic (PLEG): container finished" podID="995e7ee9-bf81-43e8-b4d8-e11603a26e0e" containerID="37ff59fda22fd24f06a226f4b6a1d27f61181aa140a8ab28edefb6d90dc6028c" exitCode=0 Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.630039 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" event={"ID":"995e7ee9-bf81-43e8-b4d8-e11603a26e0e","Type":"ContainerDied","Data":"37ff59fda22fd24f06a226f4b6a1d27f61181aa140a8ab28edefb6d90dc6028c"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.638444 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kw6xm" event={"ID":"84b7af84-2b2c-48f7-b7b0-e38fb0f7d702","Type":"ContainerStarted","Data":"8b40c2b92278e00dcd889bead29bf13ab82b4617ef1fd3f755fba660fe80034a"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.648823 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" podStartSLOduration=62.648802339 podStartE2EDuration="1m2.648802339s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.63433641 +0000 UTC m=+115.435434474" watchObservedRunningTime="2026-02-25 13:22:37.648802339 +0000 UTC m=+115.449900403" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.662068 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-2dz2z" podStartSLOduration=7.662031011 podStartE2EDuration="7.662031011s" podCreationTimestamp="2026-02-25 13:22:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.661975719 +0000 UTC m=+115.463073773" watchObservedRunningTime="2026-02-25 13:22:37.662031011 +0000 UTC m=+115.463129065" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.662792 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dqj92" event={"ID":"d6db35c3-7baa-4452-b8ff-5c20470142d6","Type":"ContainerStarted","Data":"086cff1e238bd2335dff4332eed7b2e4cbd5707ee7dede483a91862e33ed248b"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.662837 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dqj92" event={"ID":"d6db35c3-7baa-4452-b8ff-5c20470142d6","Type":"ContainerStarted","Data":"6d5bc9980c124b30e11bc3055294e3ec7133fac295d59d2bd5b6b9929bfa4718"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.668922 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" event={"ID":"1b2fb1a6-e1d1-48e0-83dc-17143971294e","Type":"ContainerStarted","Data":"3ae1c93addc3743f6a3c4f03cb5cc5f88484ac039710085e2e9e90d700a80eac"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.669685 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.680968 4815 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-6th6z container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.681023 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" podUID="1b2fb1a6-e1d1-48e0-83dc-17143971294e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.710830 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" event={"ID":"739241b1-c772-4ef7-b149-e2098da43d67","Type":"ContainerStarted","Data":"401910ba4cc6aed5b74b3e3a090a147396994b2b7269f63f02bb53ae4afb48e1"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.711937 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.715892 4815 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-4klbc container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.715946 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" podUID="60971790-c5dc-4884-ae44-e91d035cbe31" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.716030 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c699m" event={"ID":"643cac92-ba9e-4d3c-97b7-ca525f8fc046","Type":"ContainerStarted","Data":"87928e0fdac286a74b6062030dccacf9498ca1462a54ed60113c8990726bbb36"} Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.718493 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-8jnnl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" start-of-body= Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.718805 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8jnnl" podUID="78d30b59-0a66-44e5-b0b9-a2864e4d4552" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.718887 4815 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-h7bdq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:5443/healthz\": dial tcp 10.217.0.20:5443: connect: connection refused" start-of-body= Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.718932 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" podUID="739241b1-c772-4ef7-b149-e2098da43d67" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.20:5443/healthz\": dial tcp 10.217.0.20:5443: connect: connection refused" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.721436 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.722394 4815 patch_prober.go:28] interesting pod/console-operator-58897d9998-w8lg6 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.722426 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-w8lg6" podUID="bdc0843c-6110-4d69-b444-7e5384f858e1" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.726219 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:37 crc kubenswrapper[4815]: E0225 13:22:37.726790 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:38.226774687 +0000 UTC m=+116.027872821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.731243 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ff75l" podStartSLOduration=62.731232602 podStartE2EDuration="1m2.731232602s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.708748329 +0000 UTC m=+115.509846383" watchObservedRunningTime="2026-02-25 13:22:37.731232602 +0000 UTC m=+115.532330656" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.741243 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qdgb5" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.753549 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7g5t4" podStartSLOduration=61.753531689 podStartE2EDuration="1m1.753531689s" podCreationTimestamp="2026-02-25 13:21:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.730076147 +0000 UTC m=+115.531174191" watchObservedRunningTime="2026-02-25 13:22:37.753531689 +0000 UTC m=+115.554629753" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.758236 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-lttkb" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.772276 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-9t9rg" podStartSLOduration=62.772254098 podStartE2EDuration="1m2.772254098s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.753084295 +0000 UTC m=+115.554182379" watchObservedRunningTime="2026-02-25 13:22:37.772254098 +0000 UTC m=+115.573352152" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.789593 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-tswct" podStartSLOduration=62.789574603 podStartE2EDuration="1m2.789574603s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.775200268 +0000 UTC m=+115.576298332" watchObservedRunningTime="2026-02-25 13:22:37.789574603 +0000 UTC m=+115.590672657" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.790587 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-dv7nv"] Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.805785 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" podStartSLOduration=63.805767946 podStartE2EDuration="1m3.805767946s" podCreationTimestamp="2026-02-25 13:21:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.801234208 +0000 UTC m=+115.602332272" watchObservedRunningTime="2026-02-25 13:22:37.805767946 +0000 UTC m=+115.606866000" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.823095 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:37 crc kubenswrapper[4815]: E0225 13:22:37.824363 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:38.32434418 +0000 UTC m=+116.125442234 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.833604 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w2rk4" podStartSLOduration=62.833584611 podStartE2EDuration="1m2.833584611s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.815436779 +0000 UTC m=+115.616534833" watchObservedRunningTime="2026-02-25 13:22:37.833584611 +0000 UTC m=+115.634682675" Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.931217 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:37 crc kubenswrapper[4815]: E0225 13:22:37.931517 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:38.431492244 +0000 UTC m=+116.232590298 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:37 crc kubenswrapper[4815]: I0225 13:22:37.973963 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" podStartSLOduration=61.973930982 podStartE2EDuration="1m1.973930982s" podCreationTimestamp="2026-02-25 13:21:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:37.939831847 +0000 UTC m=+115.740929901" watchObservedRunningTime="2026-02-25 13:22:37.973930982 +0000 UTC m=+115.775029046" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.033090 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:38 crc kubenswrapper[4815]: E0225 13:22:38.033576 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:38.533555742 +0000 UTC m=+116.334653806 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.057501 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-wdwxg" podStartSLOduration=62.057470609 podStartE2EDuration="1m2.057470609s" podCreationTimestamp="2026-02-25 13:21:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:38.021659522 +0000 UTC m=+115.822757576" watchObservedRunningTime="2026-02-25 13:22:38.057470609 +0000 UTC m=+115.858568663" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.074342 4815 ???:1] "http: TLS handshake error from 192.168.126.11:50564: no serving certificate available for the kubelet" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.124826 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-c699m" podStartSLOduration=63.124807674 podStartE2EDuration="1m3.124807674s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:38.09404926 +0000 UTC m=+115.895147334" watchObservedRunningTime="2026-02-25 13:22:38.124807674 +0000 UTC m=+115.925905728" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.137392 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:38 crc kubenswrapper[4815]: E0225 13:22:38.137800 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:38.637784728 +0000 UTC m=+116.438882782 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.143682 4815 patch_prober.go:28] interesting pod/router-default-5444994796-5chwn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 13:22:38 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Feb 25 13:22:38 crc kubenswrapper[4815]: [+]process-running ok Feb 25 13:22:38 crc kubenswrapper[4815]: healthz check failed Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.143717 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5chwn" podUID="e9ae59fe-6c8c-4906-8c36-2b808d0474c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.147270 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-dqj92" podStartSLOduration=63.147258226 podStartE2EDuration="1m3.147258226s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:38.144186172 +0000 UTC m=+115.945284226" watchObservedRunningTime="2026-02-25 13:22:38.147258226 +0000 UTC m=+115.948356280" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.149202 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-njtsc" podStartSLOduration=63.149196154 podStartE2EDuration="1m3.149196154s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:38.126762563 +0000 UTC m=+115.927860617" watchObservedRunningTime="2026-02-25 13:22:38.149196154 +0000 UTC m=+115.950294208" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.185187 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" podStartSLOduration=63.185165796 podStartE2EDuration="1m3.185165796s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:38.18396601 +0000 UTC m=+115.985064064" watchObservedRunningTime="2026-02-25 13:22:38.185165796 +0000 UTC m=+115.986263850" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.238345 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" podStartSLOduration=63.238329311 podStartE2EDuration="1m3.238329311s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:38.23434123 +0000 UTC m=+116.035439294" watchObservedRunningTime="2026-02-25 13:22:38.238329311 +0000 UTC m=+116.039427365" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.239715 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:38 crc kubenswrapper[4815]: E0225 13:22:38.239990 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:38.739979431 +0000 UTC m=+116.541077485 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.341077 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:38 crc kubenswrapper[4815]: E0225 13:22:38.341446 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:38.841431031 +0000 UTC m=+116.642529085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.442346 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:38 crc kubenswrapper[4815]: E0225 13:22:38.442561 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:38.942532641 +0000 UTC m=+116.743630695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.442950 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:38 crc kubenswrapper[4815]: E0225 13:22:38.443241 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:38.943229283 +0000 UTC m=+116.744327337 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.544144 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:38 crc kubenswrapper[4815]: E0225 13:22:38.544477 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:39.044460397 +0000 UTC m=+116.845558451 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.645385 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:38 crc kubenswrapper[4815]: E0225 13:22:38.645700 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:39.145688101 +0000 UTC m=+116.946786155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.726102 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-j9qns" event={"ID":"3c08c2da-42cd-47ba-92e6-aed1d48855d3","Type":"ContainerStarted","Data":"2a94301c955a7a971d5f693a0bf9b8437362c1e6bc30c7ba04b69a367fd9c0e5"} Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.729213 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-shdc4" event={"ID":"a14e9f5d-f4ac-436b-8a6e-9523697d2b74","Type":"ContainerStarted","Data":"8ff1750ac0df6af2e202c126532c063f10e1ae18a97bae4051fc4e57db4c9586"} Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.734044 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" event={"ID":"995e7ee9-bf81-43e8-b4d8-e11603a26e0e","Type":"ContainerStarted","Data":"8a1fd8f467a318f58735b977683cdfd2f78cc591fb4690af79205db5fd885d25"} Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.742207 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kw6xm" event={"ID":"84b7af84-2b2c-48f7-b7b0-e38fb0f7d702","Type":"ContainerStarted","Data":"1edaed36edb106e42cdec4f548cb7f9aa04e7d972a39c6262f77519d3d21df3e"} Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.742337 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-kw6xm" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.746776 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:38 crc kubenswrapper[4815]: E0225 13:22:38.746955 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:39.246928254 +0000 UTC m=+117.048026308 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.747124 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:38 crc kubenswrapper[4815]: E0225 13:22:38.747478 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:39.247466991 +0000 UTC m=+117.048565045 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.748429 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hckbl" event={"ID":"e8a672b4-6113-4f9a-8d34-e60d9c8c40d2","Type":"ContainerStarted","Data":"c82258d3655b0c448f15b379d90d3b1ebd9031ca8cba69bc092cf33af8978dd1"} Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.752189 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gbqbs" event={"ID":"97b044d0-af42-4f42-95a8-49d6c8782246","Type":"ContainerStarted","Data":"67181af778c229e7f0af460b87da02b736d70a47530650de4dc9ecd8d852db3e"} Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.752224 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gbqbs" event={"ID":"97b044d0-af42-4f42-95a8-49d6c8782246","Type":"ContainerStarted","Data":"1b902bbb2d4acacba00fee29d6a6933a90b9363d42316b8da84b97ee71b799fb"} Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.753127 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-8jnnl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" start-of-body= Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.753160 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8jnnl" podUID="78d30b59-0a66-44e5-b0b9-a2864e4d4552" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.757244 4815 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-6th6z container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.757294 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" podUID="1b2fb1a6-e1d1-48e0-83dc-17143971294e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.772313 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-shdc4" podStartSLOduration=63.772295255 podStartE2EDuration="1m3.772295255s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:38.771404128 +0000 UTC m=+116.572502182" watchObservedRunningTime="2026-02-25 13:22:38.772295255 +0000 UTC m=+116.573393309" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.781035 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-w8lg6" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.792385 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" podStartSLOduration=63.792369455 podStartE2EDuration="1m3.792369455s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:38.789321092 +0000 UTC m=+116.590419146" watchObservedRunningTime="2026-02-25 13:22:38.792369455 +0000 UTC m=+116.593467509" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.827960 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-hckbl" podStartSLOduration=63.827944436 podStartE2EDuration="1m3.827944436s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:38.825573273 +0000 UTC m=+116.626671327" watchObservedRunningTime="2026-02-25 13:22:38.827944436 +0000 UTC m=+116.629042490" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.848716 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:38 crc kubenswrapper[4815]: E0225 13:22:38.851110 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:39.351091478 +0000 UTC m=+117.152189522 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.871706 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-gbqbs" podStartSLOduration=63.871684923 podStartE2EDuration="1m3.871684923s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:38.871329172 +0000 UTC m=+116.672427226" watchObservedRunningTime="2026-02-25 13:22:38.871684923 +0000 UTC m=+116.672782977" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.874247 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.875722 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.877037 4815 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-qbnmz container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.19:8443/livez\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.877083 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" podUID="995e7ee9-bf81-43e8-b4d8-e11603a26e0e" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.19:8443/livez\": dial tcp 10.217.0.19:8443: connect: connection refused" Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.951789 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:38 crc kubenswrapper[4815]: E0225 13:22:38.955093 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:39.455077245 +0000 UTC m=+117.256175389 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:38 crc kubenswrapper[4815]: I0225 13:22:38.974501 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-kw6xm" podStartSLOduration=8.974484665 podStartE2EDuration="8.974484665s" podCreationTimestamp="2026-02-25 13:22:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:38.938889854 +0000 UTC m=+116.739987908" watchObservedRunningTime="2026-02-25 13:22:38.974484665 +0000 UTC m=+116.775582719" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.052874 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:39 crc kubenswrapper[4815]: E0225 13:22:39.053272 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:39.553251837 +0000 UTC m=+117.354349891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.137450 4815 patch_prober.go:28] interesting pod/router-default-5444994796-5chwn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 13:22:39 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Feb 25 13:22:39 crc kubenswrapper[4815]: [+]process-running ok Feb 25 13:22:39 crc kubenswrapper[4815]: healthz check failed Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.137570 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5chwn" podUID="e9ae59fe-6c8c-4906-8c36-2b808d0474c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.155258 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:39 crc kubenswrapper[4815]: E0225 13:22:39.155657 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:39.655642685 +0000 UTC m=+117.456740739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.256171 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:39 crc kubenswrapper[4815]: E0225 13:22:39.256632 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:39.756614382 +0000 UTC m=+117.557712446 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.357623 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:39 crc kubenswrapper[4815]: E0225 13:22:39.357992 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:39.85797724 +0000 UTC m=+117.659075304 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.458140 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:39 crc kubenswrapper[4815]: E0225 13:22:39.458627 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:39.958607946 +0000 UTC m=+117.759706000 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.527910 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zm6qq"] Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.529477 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zm6qq" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.535332 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.541203 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zm6qq"] Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.560546 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:39 crc kubenswrapper[4815]: E0225 13:22:39.560891 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:40.060875001 +0000 UTC m=+117.861973065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.594805 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.636556 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.661254 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:39 crc kubenswrapper[4815]: E0225 13:22:39.661393 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:40.161370893 +0000 UTC m=+117.962468957 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.661826 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:39 crc kubenswrapper[4815]: E0225 13:22:39.662139 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:40.162128556 +0000 UTC m=+117.963226610 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.662348 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zspjb\" (UniqueName: \"kubernetes.io/projected/9c7db856-98cb-47a4-bad2-35ec76191138-kube-api-access-zspjb\") pod \"community-operators-zm6qq\" (UID: \"9c7db856-98cb-47a4-bad2-35ec76191138\") " pod="openshift-marketplace/community-operators-zm6qq" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.662433 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c7db856-98cb-47a4-bad2-35ec76191138-utilities\") pod \"community-operators-zm6qq\" (UID: \"9c7db856-98cb-47a4-bad2-35ec76191138\") " pod="openshift-marketplace/community-operators-zm6qq" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.662467 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c7db856-98cb-47a4-bad2-35ec76191138-catalog-content\") pod \"community-operators-zm6qq\" (UID: \"9c7db856-98cb-47a4-bad2-35ec76191138\") " pod="openshift-marketplace/community-operators-zm6qq" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.693867 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vzljd"] Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.694862 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vzljd" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.703944 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.754602 4815 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-h7bdq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.754671 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" podUID="739241b1-c772-4ef7-b149-e2098da43d67" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.20:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.763035 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.763320 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlqlg\" (UniqueName: \"kubernetes.io/projected/c0706f0a-7f64-47d9-b3be-afe1d760603f-kube-api-access-rlqlg\") pod \"certified-operators-vzljd\" (UID: \"c0706f0a-7f64-47d9-b3be-afe1d760603f\") " pod="openshift-marketplace/certified-operators-vzljd" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.763362 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0706f0a-7f64-47d9-b3be-afe1d760603f-catalog-content\") pod \"certified-operators-vzljd\" (UID: \"c0706f0a-7f64-47d9-b3be-afe1d760603f\") " pod="openshift-marketplace/certified-operators-vzljd" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.763422 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zspjb\" (UniqueName: \"kubernetes.io/projected/9c7db856-98cb-47a4-bad2-35ec76191138-kube-api-access-zspjb\") pod \"community-operators-zm6qq\" (UID: \"9c7db856-98cb-47a4-bad2-35ec76191138\") " pod="openshift-marketplace/community-operators-zm6qq" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.763454 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c7db856-98cb-47a4-bad2-35ec76191138-utilities\") pod \"community-operators-zm6qq\" (UID: \"9c7db856-98cb-47a4-bad2-35ec76191138\") " pod="openshift-marketplace/community-operators-zm6qq" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.763480 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c7db856-98cb-47a4-bad2-35ec76191138-catalog-content\") pod \"community-operators-zm6qq\" (UID: \"9c7db856-98cb-47a4-bad2-35ec76191138\") " pod="openshift-marketplace/community-operators-zm6qq" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.763608 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0706f0a-7f64-47d9-b3be-afe1d760603f-utilities\") pod \"certified-operators-vzljd\" (UID: \"c0706f0a-7f64-47d9-b3be-afe1d760603f\") " pod="openshift-marketplace/certified-operators-vzljd" Feb 25 13:22:39 crc kubenswrapper[4815]: E0225 13:22:39.763722 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:40.26370403 +0000 UTC m=+118.064802084 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.764382 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c7db856-98cb-47a4-bad2-35ec76191138-utilities\") pod \"community-operators-zm6qq\" (UID: \"9c7db856-98cb-47a4-bad2-35ec76191138\") " pod="openshift-marketplace/community-operators-zm6qq" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.764686 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c7db856-98cb-47a4-bad2-35ec76191138-catalog-content\") pod \"community-operators-zm6qq\" (UID: \"9c7db856-98cb-47a4-bad2-35ec76191138\") " pod="openshift-marketplace/community-operators-zm6qq" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.782561 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vzljd"] Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.797525 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zspjb\" (UniqueName: \"kubernetes.io/projected/9c7db856-98cb-47a4-bad2-35ec76191138-kube-api-access-zspjb\") pod \"community-operators-zm6qq\" (UID: \"9c7db856-98cb-47a4-bad2-35ec76191138\") " pod="openshift-marketplace/community-operators-zm6qq" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.805707 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" podUID="65dfccf6-aa91-41fa-95d6-fd4778065408" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9" gracePeriod=30 Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.815805 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.857790 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zm6qq" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.866285 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.866370 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.866602 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.866644 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0706f0a-7f64-47d9-b3be-afe1d760603f-utilities\") pod \"certified-operators-vzljd\" (UID: \"c0706f0a-7f64-47d9-b3be-afe1d760603f\") " pod="openshift-marketplace/certified-operators-vzljd" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.866710 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.866745 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlqlg\" (UniqueName: \"kubernetes.io/projected/c0706f0a-7f64-47d9-b3be-afe1d760603f-kube-api-access-rlqlg\") pod \"certified-operators-vzljd\" (UID: \"c0706f0a-7f64-47d9-b3be-afe1d760603f\") " pod="openshift-marketplace/certified-operators-vzljd" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.866786 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0706f0a-7f64-47d9-b3be-afe1d760603f-catalog-content\") pod \"certified-operators-vzljd\" (UID: \"c0706f0a-7f64-47d9-b3be-afe1d760603f\") " pod="openshift-marketplace/certified-operators-vzljd" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.867038 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.869161 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0706f0a-7f64-47d9-b3be-afe1d760603f-catalog-content\") pod \"certified-operators-vzljd\" (UID: \"c0706f0a-7f64-47d9-b3be-afe1d760603f\") " pod="openshift-marketplace/certified-operators-vzljd" Feb 25 13:22:39 crc kubenswrapper[4815]: E0225 13:22:39.901390 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:40.401374151 +0000 UTC m=+118.202472205 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.905566 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.916237 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0706f0a-7f64-47d9-b3be-afe1d760603f-utilities\") pod \"certified-operators-vzljd\" (UID: \"c0706f0a-7f64-47d9-b3be-afe1d760603f\") " pod="openshift-marketplace/certified-operators-vzljd" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.926550 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bkbjm"] Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.936680 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.938921 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.939394 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlqlg\" (UniqueName: \"kubernetes.io/projected/c0706f0a-7f64-47d9-b3be-afe1d760603f-kube-api-access-rlqlg\") pod \"certified-operators-vzljd\" (UID: \"c0706f0a-7f64-47d9-b3be-afe1d760603f\") " pod="openshift-marketplace/certified-operators-vzljd" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.941498 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bkbjm" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.966641 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.970354 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bkbjm"] Feb 25 13:22:39 crc kubenswrapper[4815]: I0225 13:22:39.971260 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:39 crc kubenswrapper[4815]: E0225 13:22:39.971572 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:40.471554842 +0000 UTC m=+118.272652896 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.025608 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vzljd" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.073327 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.073394 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwc99\" (UniqueName: \"kubernetes.io/projected/23366473-92d4-4958-bd1c-b6a2fb188502-kube-api-access-xwc99\") pod \"community-operators-bkbjm\" (UID: \"23366473-92d4-4958-bd1c-b6a2fb188502\") " pod="openshift-marketplace/community-operators-bkbjm" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.073426 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23366473-92d4-4958-bd1c-b6a2fb188502-catalog-content\") pod \"community-operators-bkbjm\" (UID: \"23366473-92d4-4958-bd1c-b6a2fb188502\") " pod="openshift-marketplace/community-operators-bkbjm" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.073453 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23366473-92d4-4958-bd1c-b6a2fb188502-utilities\") pod \"community-operators-bkbjm\" (UID: \"23366473-92d4-4958-bd1c-b6a2fb188502\") " pod="openshift-marketplace/community-operators-bkbjm" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.073613 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 13:22:40 crc kubenswrapper[4815]: E0225 13:22:40.073785 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:40.573770606 +0000 UTC m=+118.374868660 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.079178 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.083079 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.097352 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qffvd"] Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.098341 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qffvd" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.121613 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qffvd"] Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.149387 4815 patch_prober.go:28] interesting pod/router-default-5444994796-5chwn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 13:22:40 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Feb 25 13:22:40 crc kubenswrapper[4815]: [+]process-running ok Feb 25 13:22:40 crc kubenswrapper[4815]: healthz check failed Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.149436 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5chwn" podUID="e9ae59fe-6c8c-4906-8c36-2b808d0474c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.175114 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:40 crc kubenswrapper[4815]: E0225 13:22:40.175289 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:40.675262847 +0000 UTC m=+118.476360891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.175766 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvrgb\" (UniqueName: \"kubernetes.io/projected/2eed2795-5cbe-4dab-9e12-c9592f94befb-kube-api-access-rvrgb\") pod \"certified-operators-qffvd\" (UID: \"2eed2795-5cbe-4dab-9e12-c9592f94befb\") " pod="openshift-marketplace/certified-operators-qffvd" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.175851 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.175887 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2eed2795-5cbe-4dab-9e12-c9592f94befb-catalog-content\") pod \"certified-operators-qffvd\" (UID: \"2eed2795-5cbe-4dab-9e12-c9592f94befb\") " pod="openshift-marketplace/certified-operators-qffvd" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.176178 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwc99\" (UniqueName: \"kubernetes.io/projected/23366473-92d4-4958-bd1c-b6a2fb188502-kube-api-access-xwc99\") pod \"community-operators-bkbjm\" (UID: \"23366473-92d4-4958-bd1c-b6a2fb188502\") " pod="openshift-marketplace/community-operators-bkbjm" Feb 25 13:22:40 crc kubenswrapper[4815]: E0225 13:22:40.176214 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:40.676207086 +0000 UTC m=+118.477305140 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.176222 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23366473-92d4-4958-bd1c-b6a2fb188502-catalog-content\") pod \"community-operators-bkbjm\" (UID: \"23366473-92d4-4958-bd1c-b6a2fb188502\") " pod="openshift-marketplace/community-operators-bkbjm" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.176279 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2eed2795-5cbe-4dab-9e12-c9592f94befb-utilities\") pod \"certified-operators-qffvd\" (UID: \"2eed2795-5cbe-4dab-9e12-c9592f94befb\") " pod="openshift-marketplace/certified-operators-qffvd" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.176315 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23366473-92d4-4958-bd1c-b6a2fb188502-utilities\") pod \"community-operators-bkbjm\" (UID: \"23366473-92d4-4958-bd1c-b6a2fb188502\") " pod="openshift-marketplace/community-operators-bkbjm" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.176944 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23366473-92d4-4958-bd1c-b6a2fb188502-utilities\") pod \"community-operators-bkbjm\" (UID: \"23366473-92d4-4958-bd1c-b6a2fb188502\") " pod="openshift-marketplace/community-operators-bkbjm" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.177015 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23366473-92d4-4958-bd1c-b6a2fb188502-catalog-content\") pod \"community-operators-bkbjm\" (UID: \"23366473-92d4-4958-bd1c-b6a2fb188502\") " pod="openshift-marketplace/community-operators-bkbjm" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.245376 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwc99\" (UniqueName: \"kubernetes.io/projected/23366473-92d4-4958-bd1c-b6a2fb188502-kube-api-access-xwc99\") pod \"community-operators-bkbjm\" (UID: \"23366473-92d4-4958-bd1c-b6a2fb188502\") " pod="openshift-marketplace/community-operators-bkbjm" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.277340 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:40 crc kubenswrapper[4815]: E0225 13:22:40.277496 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:40.777479402 +0000 UTC m=+118.578577456 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.277640 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.277683 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2eed2795-5cbe-4dab-9e12-c9592f94befb-catalog-content\") pod \"certified-operators-qffvd\" (UID: \"2eed2795-5cbe-4dab-9e12-c9592f94befb\") " pod="openshift-marketplace/certified-operators-qffvd" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.277719 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2eed2795-5cbe-4dab-9e12-c9592f94befb-utilities\") pod \"certified-operators-qffvd\" (UID: \"2eed2795-5cbe-4dab-9e12-c9592f94befb\") " pod="openshift-marketplace/certified-operators-qffvd" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.277760 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvrgb\" (UniqueName: \"kubernetes.io/projected/2eed2795-5cbe-4dab-9e12-c9592f94befb-kube-api-access-rvrgb\") pod \"certified-operators-qffvd\" (UID: \"2eed2795-5cbe-4dab-9e12-c9592f94befb\") " pod="openshift-marketplace/certified-operators-qffvd" Feb 25 13:22:40 crc kubenswrapper[4815]: E0225 13:22:40.277945 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:40.777935046 +0000 UTC m=+118.579033100 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.278667 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2eed2795-5cbe-4dab-9e12-c9592f94befb-catalog-content\") pod \"certified-operators-qffvd\" (UID: \"2eed2795-5cbe-4dab-9e12-c9592f94befb\") " pod="openshift-marketplace/certified-operators-qffvd" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.278698 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2eed2795-5cbe-4dab-9e12-c9592f94befb-utilities\") pod \"certified-operators-qffvd\" (UID: \"2eed2795-5cbe-4dab-9e12-c9592f94befb\") " pod="openshift-marketplace/certified-operators-qffvd" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.323268 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvrgb\" (UniqueName: \"kubernetes.io/projected/2eed2795-5cbe-4dab-9e12-c9592f94befb-kube-api-access-rvrgb\") pod \"certified-operators-qffvd\" (UID: \"2eed2795-5cbe-4dab-9e12-c9592f94befb\") " pod="openshift-marketplace/certified-operators-qffvd" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.324151 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bkbjm" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.325989 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h7bdq" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.363847 4815 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-z7rhl container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.363887 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" podUID="5358560f-45f9-4548-8521-fad61a51f5a9" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.364160 4815 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-z7rhl container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.364180 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" podUID="5358560f-45f9-4548-8521-fad61a51f5a9" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.380770 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:40 crc kubenswrapper[4815]: E0225 13:22:40.381133 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:40.881118359 +0000 UTC m=+118.682216413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.381154 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4klbc"] Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.381311 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" podUID="60971790-c5dc-4884-ae44-e91d035cbe31" containerName="controller-manager" containerID="cri-o://8bd4854abbc245cd236b0340b06154edf75ebec05eb8c528b58affe3c748103b" gracePeriod=30 Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.409760 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.417617 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qffvd" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.461781 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vzljd"] Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.472978 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd"] Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.482347 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:40 crc kubenswrapper[4815]: E0225 13:22:40.482709 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:40.982694993 +0000 UTC m=+118.783793047 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.586679 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:40 crc kubenswrapper[4815]: E0225 13:22:40.587021 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:41.08700594 +0000 UTC m=+118.888103994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.599328 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zm6qq"] Feb 25 13:22:40 crc kubenswrapper[4815]: W0225 13:22:40.646616 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c7db856_98cb_47a4_bad2_35ec76191138.slice/crio-96628b7b501006f4311cedab3503ef1041488bc91b51ffa7b2d8727c9c43b7de WatchSource:0}: Error finding container 96628b7b501006f4311cedab3503ef1041488bc91b51ffa7b2d8727c9c43b7de: Status 404 returned error can't find the container with id 96628b7b501006f4311cedab3503ef1041488bc91b51ffa7b2d8727c9c43b7de Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.667673 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.668393 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.674902 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.675009 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.676772 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.679941 4815 ???:1] "http: TLS handshake error from 192.168.126.11:59468: no serving certificate available for the kubelet" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.687880 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c846b62d-3716-43ac-8f8d-b68c593f3b49-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c846b62d-3716-43ac-8f8d-b68c593f3b49\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.687930 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.687993 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c846b62d-3716-43ac-8f8d-b68c593f3b49-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c846b62d-3716-43ac-8f8d-b68c593f3b49\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 13:22:40 crc kubenswrapper[4815]: E0225 13:22:40.688277 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:41.188263055 +0000 UTC m=+118.989361109 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.788902 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:40 crc kubenswrapper[4815]: E0225 13:22:40.789005 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:41.288986734 +0000 UTC m=+119.090084788 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.789320 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c846b62d-3716-43ac-8f8d-b68c593f3b49-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c846b62d-3716-43ac-8f8d-b68c593f3b49\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.789364 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c846b62d-3716-43ac-8f8d-b68c593f3b49-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c846b62d-3716-43ac-8f8d-b68c593f3b49\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.789388 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.789577 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c846b62d-3716-43ac-8f8d-b68c593f3b49-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c846b62d-3716-43ac-8f8d-b68c593f3b49\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 13:22:40 crc kubenswrapper[4815]: E0225 13:22:40.789861 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:41.28983919 +0000 UTC m=+119.090937274 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.821570 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzljd" event={"ID":"c0706f0a-7f64-47d9-b3be-afe1d760603f","Type":"ContainerStarted","Data":"c64f52522b823598007d99057217b63dc09189b0ca765150b4e30b4eb73180e1"} Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.824200 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zm6qq" event={"ID":"9c7db856-98cb-47a4-bad2-35ec76191138","Type":"ContainerStarted","Data":"96628b7b501006f4311cedab3503ef1041488bc91b51ffa7b2d8727c9c43b7de"} Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.840834 4815 generic.go:334] "Generic (PLEG): container finished" podID="80629dad-7d0a-46b3-bb52-09b75e7dfeb9" containerID="5fcc6127de7c4f26b8ab72ffdab14f84fee434383c588d93cc0b01819e27c9cb" exitCode=0 Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.840915 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" event={"ID":"80629dad-7d0a-46b3-bb52-09b75e7dfeb9","Type":"ContainerDied","Data":"5fcc6127de7c4f26b8ab72ffdab14f84fee434383c588d93cc0b01819e27c9cb"} Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.855280 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-j9qns" event={"ID":"3c08c2da-42cd-47ba-92e6-aed1d48855d3","Type":"ContainerStarted","Data":"d57f8f34bf7ddce92039c4feb02d0c467bcfea7a2fed5e1d9c74d9e2a553fd31"} Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.871524 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c846b62d-3716-43ac-8f8d-b68c593f3b49-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c846b62d-3716-43ac-8f8d-b68c593f3b49\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.890778 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:40 crc kubenswrapper[4815]: E0225 13:22:40.891087 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:41.391072614 +0000 UTC m=+119.192170668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.919315 4815 generic.go:334] "Generic (PLEG): container finished" podID="60971790-c5dc-4884-ae44-e91d035cbe31" containerID="8bd4854abbc245cd236b0340b06154edf75ebec05eb8c528b58affe3c748103b" exitCode=0 Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.919740 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" event={"ID":"60971790-c5dc-4884-ae44-e91d035cbe31","Type":"ContainerDied","Data":"8bd4854abbc245cd236b0340b06154edf75ebec05eb8c528b58affe3c748103b"} Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.920519 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" podUID="72d966ad-816d-4483-b9e6-44edb10f9d75" containerName="route-controller-manager" containerID="cri-o://91dfc815ac31db022e98712effb6e15a11dfc06a27f450d84b1c5fa9c3d9e40a" gracePeriod=30 Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.925162 4815 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 25 13:22:40 crc kubenswrapper[4815]: W0225 13:22:40.943531 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-d6576814a32a8a7bf3edc43851227820e7aa0135cd05414a25da52b0309eaf20 WatchSource:0}: Error finding container d6576814a32a8a7bf3edc43851227820e7aa0135cd05414a25da52b0309eaf20: Status 404 returned error can't find the container with id d6576814a32a8a7bf3edc43851227820e7aa0135cd05414a25da52b0309eaf20 Feb 25 13:22:40 crc kubenswrapper[4815]: I0225 13:22:40.998017 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:41 crc kubenswrapper[4815]: E0225 13:22:40.999843 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 13:22:41.499829807 +0000 UTC m=+119.300927961 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-z4lg2" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.001380 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.100377 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:41 crc kubenswrapper[4815]: E0225 13:22:41.100657 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 13:22:41.600639598 +0000 UTC m=+119.401737652 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.112753 4815 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-25T13:22:40.925183719Z","Handler":null,"Name":""} Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.124157 4815 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.124192 4815 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.147744 4815 patch_prober.go:28] interesting pod/router-default-5444994796-5chwn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 13:22:41 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Feb 25 13:22:41 crc kubenswrapper[4815]: [+]process-running ok Feb 25 13:22:41 crc kubenswrapper[4815]: healthz check failed Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.147799 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5chwn" podUID="e9ae59fe-6c8c-4906-8c36-2b808d0474c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.188990 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qffvd"] Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.203293 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.211105 4815 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.211149 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.231732 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:41 crc kubenswrapper[4815]: W0225 13:22:41.260213 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-65d5cec476a919a81ab1e878b10a9481c0cc92f7c816daf12101d975f30979e5 WatchSource:0}: Error finding container 65d5cec476a919a81ab1e878b10a9481c0cc92f7c816daf12101d975f30979e5: Status 404 returned error can't find the container with id 65d5cec476a919a81ab1e878b10a9481c0cc92f7c816daf12101d975f30979e5 Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.270192 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-z4lg2\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.304888 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d6t6\" (UniqueName: \"kubernetes.io/projected/60971790-c5dc-4884-ae44-e91d035cbe31-kube-api-access-2d6t6\") pod \"60971790-c5dc-4884-ae44-e91d035cbe31\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.304932 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60971790-c5dc-4884-ae44-e91d035cbe31-serving-cert\") pod \"60971790-c5dc-4884-ae44-e91d035cbe31\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.305021 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-client-ca\") pod \"60971790-c5dc-4884-ae44-e91d035cbe31\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.305047 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-proxy-ca-bundles\") pod \"60971790-c5dc-4884-ae44-e91d035cbe31\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.305129 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.305151 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-config\") pod \"60971790-c5dc-4884-ae44-e91d035cbe31\" (UID: \"60971790-c5dc-4884-ae44-e91d035cbe31\") " Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.306330 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-config" (OuterVolumeSpecName: "config") pod "60971790-c5dc-4884-ae44-e91d035cbe31" (UID: "60971790-c5dc-4884-ae44-e91d035cbe31"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.306682 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.306740 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "60971790-c5dc-4884-ae44-e91d035cbe31" (UID: "60971790-c5dc-4884-ae44-e91d035cbe31"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.308347 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-client-ca" (OuterVolumeSpecName: "client-ca") pod "60971790-c5dc-4884-ae44-e91d035cbe31" (UID: "60971790-c5dc-4884-ae44-e91d035cbe31"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.312539 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60971790-c5dc-4884-ae44-e91d035cbe31-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "60971790-c5dc-4884-ae44-e91d035cbe31" (UID: "60971790-c5dc-4884-ae44-e91d035cbe31"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.314518 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60971790-c5dc-4884-ae44-e91d035cbe31-kube-api-access-2d6t6" (OuterVolumeSpecName: "kube-api-access-2d6t6") pod "60971790-c5dc-4884-ae44-e91d035cbe31" (UID: "60971790-c5dc-4884-ae44-e91d035cbe31"). InnerVolumeSpecName "kube-api-access-2d6t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.320878 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.354465 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bkbjm"] Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.406130 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.406675 4815 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.406686 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60971790-c5dc-4884-ae44-e91d035cbe31-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.406695 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d6t6\" (UniqueName: \"kubernetes.io/projected/60971790-c5dc-4884-ae44-e91d035cbe31-kube-api-access-2d6t6\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.406705 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60971790-c5dc-4884-ae44-e91d035cbe31-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.407688 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 25 13:22:41 crc kubenswrapper[4815]: W0225 13:22:41.444958 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podc846b62d_3716_43ac_8f8d_b68c593f3b49.slice/crio-dd1540c82e4c070ffe35b6e62f4def0f166b15f46cd1b7c78221300c51d0cc55 WatchSource:0}: Error finding container dd1540c82e4c070ffe35b6e62f4def0f166b15f46cd1b7c78221300c51d0cc55: Status 404 returned error can't find the container with id dd1540c82e4c070ffe35b6e62f4def0f166b15f46cd1b7c78221300c51d0cc55 Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.461115 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.511400 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72d966ad-816d-4483-b9e6-44edb10f9d75-config\") pod \"72d966ad-816d-4483-b9e6-44edb10f9d75\" (UID: \"72d966ad-816d-4483-b9e6-44edb10f9d75\") " Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.511620 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/72d966ad-816d-4483-b9e6-44edb10f9d75-client-ca\") pod \"72d966ad-816d-4483-b9e6-44edb10f9d75\" (UID: \"72d966ad-816d-4483-b9e6-44edb10f9d75\") " Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.511648 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brbmt\" (UniqueName: \"kubernetes.io/projected/72d966ad-816d-4483-b9e6-44edb10f9d75-kube-api-access-brbmt\") pod \"72d966ad-816d-4483-b9e6-44edb10f9d75\" (UID: \"72d966ad-816d-4483-b9e6-44edb10f9d75\") " Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.511690 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72d966ad-816d-4483-b9e6-44edb10f9d75-serving-cert\") pod \"72d966ad-816d-4483-b9e6-44edb10f9d75\" (UID: \"72d966ad-816d-4483-b9e6-44edb10f9d75\") " Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.515268 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72d966ad-816d-4483-b9e6-44edb10f9d75-client-ca" (OuterVolumeSpecName: "client-ca") pod "72d966ad-816d-4483-b9e6-44edb10f9d75" (UID: "72d966ad-816d-4483-b9e6-44edb10f9d75"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.515775 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72d966ad-816d-4483-b9e6-44edb10f9d75-config" (OuterVolumeSpecName: "config") pod "72d966ad-816d-4483-b9e6-44edb10f9d75" (UID: "72d966ad-816d-4483-b9e6-44edb10f9d75"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.526145 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72d966ad-816d-4483-b9e6-44edb10f9d75-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "72d966ad-816d-4483-b9e6-44edb10f9d75" (UID: "72d966ad-816d-4483-b9e6-44edb10f9d75"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.530979 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72d966ad-816d-4483-b9e6-44edb10f9d75-kube-api-access-brbmt" (OuterVolumeSpecName: "kube-api-access-brbmt") pod "72d966ad-816d-4483-b9e6-44edb10f9d75" (UID: "72d966ad-816d-4483-b9e6-44edb10f9d75"). InnerVolumeSpecName "kube-api-access-brbmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.613204 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/72d966ad-816d-4483-b9e6-44edb10f9d75-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.613245 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brbmt\" (UniqueName: \"kubernetes.io/projected/72d966ad-816d-4483-b9e6-44edb10f9d75-kube-api-access-brbmt\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.613258 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72d966ad-816d-4483-b9e6-44edb10f9d75-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.613270 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72d966ad-816d-4483-b9e6-44edb10f9d75-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.683028 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x6zz5"] Feb 25 13:22:41 crc kubenswrapper[4815]: E0225 13:22:41.683287 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72d966ad-816d-4483-b9e6-44edb10f9d75" containerName="route-controller-manager" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.683303 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="72d966ad-816d-4483-b9e6-44edb10f9d75" containerName="route-controller-manager" Feb 25 13:22:41 crc kubenswrapper[4815]: E0225 13:22:41.683334 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60971790-c5dc-4884-ae44-e91d035cbe31" containerName="controller-manager" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.683340 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="60971790-c5dc-4884-ae44-e91d035cbe31" containerName="controller-manager" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.683487 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="72d966ad-816d-4483-b9e6-44edb10f9d75" containerName="route-controller-manager" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.683565 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="60971790-c5dc-4884-ae44-e91d035cbe31" containerName="controller-manager" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.684858 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6zz5" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.692456 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.695407 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6zz5"] Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.714712 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-utilities\") pod \"redhat-marketplace-x6zz5\" (UID: \"6d00f8b4-1bf6-4a47-9beb-2226f26d695b\") " pod="openshift-marketplace/redhat-marketplace-x6zz5" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.714787 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw8ct\" (UniqueName: \"kubernetes.io/projected/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-kube-api-access-fw8ct\") pod \"redhat-marketplace-x6zz5\" (UID: \"6d00f8b4-1bf6-4a47-9beb-2226f26d695b\") " pod="openshift-marketplace/redhat-marketplace-x6zz5" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.714943 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-catalog-content\") pod \"redhat-marketplace-x6zz5\" (UID: \"6d00f8b4-1bf6-4a47-9beb-2226f26d695b\") " pod="openshift-marketplace/redhat-marketplace-x6zz5" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.816793 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-utilities\") pod \"redhat-marketplace-x6zz5\" (UID: \"6d00f8b4-1bf6-4a47-9beb-2226f26d695b\") " pod="openshift-marketplace/redhat-marketplace-x6zz5" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.816850 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw8ct\" (UniqueName: \"kubernetes.io/projected/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-kube-api-access-fw8ct\") pod \"redhat-marketplace-x6zz5\" (UID: \"6d00f8b4-1bf6-4a47-9beb-2226f26d695b\") " pod="openshift-marketplace/redhat-marketplace-x6zz5" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.816930 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-catalog-content\") pod \"redhat-marketplace-x6zz5\" (UID: \"6d00f8b4-1bf6-4a47-9beb-2226f26d695b\") " pod="openshift-marketplace/redhat-marketplace-x6zz5" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.817764 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-utilities\") pod \"redhat-marketplace-x6zz5\" (UID: \"6d00f8b4-1bf6-4a47-9beb-2226f26d695b\") " pod="openshift-marketplace/redhat-marketplace-x6zz5" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.817808 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-catalog-content\") pod \"redhat-marketplace-x6zz5\" (UID: \"6d00f8b4-1bf6-4a47-9beb-2226f26d695b\") " pod="openshift-marketplace/redhat-marketplace-x6zz5" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.836662 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw8ct\" (UniqueName: \"kubernetes.io/projected/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-kube-api-access-fw8ct\") pod \"redhat-marketplace-x6zz5\" (UID: \"6d00f8b4-1bf6-4a47-9beb-2226f26d695b\") " pod="openshift-marketplace/redhat-marketplace-x6zz5" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.885351 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-z4lg2"] Feb 25 13:22:41 crc kubenswrapper[4815]: W0225 13:22:41.895404 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod072fbe81_88e1_4755_bb59_5ee95d72760f.slice/crio-c7ff748a7a9ef0f49190a50fd5aac69c46afab5e9cc542eee8b6efcf1420a0d6 WatchSource:0}: Error finding container c7ff748a7a9ef0f49190a50fd5aac69c46afab5e9cc542eee8b6efcf1420a0d6: Status 404 returned error can't find the container with id c7ff748a7a9ef0f49190a50fd5aac69c46afab5e9cc542eee8b6efcf1420a0d6 Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.931651 4815 generic.go:334] "Generic (PLEG): container finished" podID="23366473-92d4-4958-bd1c-b6a2fb188502" containerID="37b745cb4e1580b68d2ba3d751e25df91a9fe3dee311b410061c12dd30ff8026" exitCode=0 Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.931730 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkbjm" event={"ID":"23366473-92d4-4958-bd1c-b6a2fb188502","Type":"ContainerDied","Data":"37b745cb4e1580b68d2ba3d751e25df91a9fe3dee311b410061c12dd30ff8026"} Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.931762 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkbjm" event={"ID":"23366473-92d4-4958-bd1c-b6a2fb188502","Type":"ContainerStarted","Data":"b4fb42e181cacc2765330219221d0bf5a53a5c1eede8dc6bfd0dc88ede6f5f8d"} Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.933288 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" event={"ID":"072fbe81-88e1-4755-bb59-5ee95d72760f","Type":"ContainerStarted","Data":"c7ff748a7a9ef0f49190a50fd5aac69c46afab5e9cc542eee8b6efcf1420a0d6"} Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.935248 4815 generic.go:334] "Generic (PLEG): container finished" podID="9c7db856-98cb-47a4-bad2-35ec76191138" containerID="a2379a4ffbdde16378a20d3074a26486558618bb4fca5b20693493f9d889e589" exitCode=0 Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.935308 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zm6qq" event={"ID":"9c7db856-98cb-47a4-bad2-35ec76191138","Type":"ContainerDied","Data":"a2379a4ffbdde16378a20d3074a26486558618bb4fca5b20693493f9d889e589"} Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.940449 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"c5c2ecd2f1487adc56800c0d1f7dc7a829c7973830f4d4782a94fa078ccb1086"} Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.940522 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"65d5cec476a919a81ab1e878b10a9481c0cc92f7c816daf12101d975f30979e5"} Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.942899 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4c8d245b601c59036c7b77187d2b52e63b2e1fe569216c50224b74a340568591"} Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.942962 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"c304196399c5c2b1a7efb8835821a0b5554e2aa38e53587b2cd734f7e8b04895"} Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.950291 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-j9qns" event={"ID":"3c08c2da-42cd-47ba-92e6-aed1d48855d3","Type":"ContainerStarted","Data":"5117d9ccbf9f56514dbd074447330330d3419f4ac27ba11cda6a8bfcb798a448"} Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.950334 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-j9qns" event={"ID":"3c08c2da-42cd-47ba-92e6-aed1d48855d3","Type":"ContainerStarted","Data":"863b3dbb3af76a8e65b32956d1fda5311ba78674cdc28a110099688ee6cde380"} Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.955420 4815 generic.go:334] "Generic (PLEG): container finished" podID="c0706f0a-7f64-47d9-b3be-afe1d760603f" containerID="bd14764d84cfe21d3b0123f56abe3441bbb9f54618dcc498b0c3c007bfa74a03" exitCode=0 Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.955491 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzljd" event={"ID":"c0706f0a-7f64-47d9-b3be-afe1d760603f","Type":"ContainerDied","Data":"bd14764d84cfe21d3b0123f56abe3441bbb9f54618dcc498b0c3c007bfa74a03"} Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.968371 4815 generic.go:334] "Generic (PLEG): container finished" podID="2eed2795-5cbe-4dab-9e12-c9592f94befb" containerID="5c55e82bf68f55bf6a6aa089e5a131267568f93fed4e8cd6c8782d2bf23a26ab" exitCode=0 Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.968467 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qffvd" event={"ID":"2eed2795-5cbe-4dab-9e12-c9592f94befb","Type":"ContainerDied","Data":"5c55e82bf68f55bf6a6aa089e5a131267568f93fed4e8cd6c8782d2bf23a26ab"} Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.968546 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qffvd" event={"ID":"2eed2795-5cbe-4dab-9e12-c9592f94befb","Type":"ContainerStarted","Data":"67ba65baa8bfafdbb506dfc3eea5e0feb3ae3a47a7e21eed55d24585282cc6aa"} Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.978462 4815 generic.go:334] "Generic (PLEG): container finished" podID="72d966ad-816d-4483-b9e6-44edb10f9d75" containerID="91dfc815ac31db022e98712effb6e15a11dfc06a27f450d84b1c5fa9c3d9e40a" exitCode=0 Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.978546 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" event={"ID":"72d966ad-816d-4483-b9e6-44edb10f9d75","Type":"ContainerDied","Data":"91dfc815ac31db022e98712effb6e15a11dfc06a27f450d84b1c5fa9c3d9e40a"} Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.978579 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" event={"ID":"72d966ad-816d-4483-b9e6-44edb10f9d75","Type":"ContainerDied","Data":"08c83bfc91ff7986916e4fdc443458a7467f3928ad09a3629eb141adebe663f2"} Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.978577 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.978601 4815 scope.go:117] "RemoveContainer" containerID="91dfc815ac31db022e98712effb6e15a11dfc06a27f450d84b1c5fa9c3d9e40a" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.988996 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" Feb 25 13:22:41 crc kubenswrapper[4815]: I0225 13:22:41.991752 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4klbc" event={"ID":"60971790-c5dc-4884-ae44-e91d035cbe31","Type":"ContainerDied","Data":"b3426afcad311531ab9fc6fbdc97435f2d4bfdca83a6a49fae81620d68f1cdae"} Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.000806 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"07a05e13ac6f620339f87bc33bb8a8e0e02259711fdeb0afea96d768c23a2a45"} Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.000858 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d6576814a32a8a7bf3edc43851227820e7aa0135cd05414a25da52b0309eaf20"} Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.001534 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.006151 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c846b62d-3716-43ac-8f8d-b68c593f3b49","Type":"ContainerStarted","Data":"d77bc3984367ebd298a4511fccf39ef8b8d4ddbc459782d2dddcd1a28f773bc5"} Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.006194 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c846b62d-3716-43ac-8f8d-b68c593f3b49","Type":"ContainerStarted","Data":"dd1540c82e4c070ffe35b6e62f4def0f166b15f46cd1b7c78221300c51d0cc55"} Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.009422 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6zz5" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.014581 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.015206 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.017732 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.018457 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.018730 4815 scope.go:117] "RemoveContainer" containerID="91dfc815ac31db022e98712effb6e15a11dfc06a27f450d84b1c5fa9c3d9e40a" Feb 25 13:22:42 crc kubenswrapper[4815]: E0225 13:22:42.022678 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91dfc815ac31db022e98712effb6e15a11dfc06a27f450d84b1c5fa9c3d9e40a\": container with ID starting with 91dfc815ac31db022e98712effb6e15a11dfc06a27f450d84b1c5fa9c3d9e40a not found: ID does not exist" containerID="91dfc815ac31db022e98712effb6e15a11dfc06a27f450d84b1c5fa9c3d9e40a" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.022714 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91dfc815ac31db022e98712effb6e15a11dfc06a27f450d84b1c5fa9c3d9e40a"} err="failed to get container status \"91dfc815ac31db022e98712effb6e15a11dfc06a27f450d84b1c5fa9c3d9e40a\": rpc error: code = NotFound desc = could not find container \"91dfc815ac31db022e98712effb6e15a11dfc06a27f450d84b1c5fa9c3d9e40a\": container with ID starting with 91dfc815ac31db022e98712effb6e15a11dfc06a27f450d84b1c5fa9c3d9e40a not found: ID does not exist" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.022834 4815 scope.go:117] "RemoveContainer" containerID="8bd4854abbc245cd236b0340b06154edf75ebec05eb8c528b58affe3c748103b" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.041630 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.052310 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-j9qns" podStartSLOduration=12.052289946 podStartE2EDuration="12.052289946s" podCreationTimestamp="2026-02-25 13:22:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:42.033658399 +0000 UTC m=+119.834756453" watchObservedRunningTime="2026-02-25 13:22:42.052289946 +0000 UTC m=+119.853388000" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.096675 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-96mhl"] Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.097646 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96mhl" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.116615 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-96mhl"] Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.121462 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd"] Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.123838 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/931648f3-732f-4bb1-81d7-6c71f2bf2971-catalog-content\") pod \"redhat-marketplace-96mhl\" (UID: \"931648f3-732f-4bb1-81d7-6c71f2bf2971\") " pod="openshift-marketplace/redhat-marketplace-96mhl" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.124703 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b43143bf-1e68-44ef-ac87-0c4936f540bd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b43143bf-1e68-44ef-ac87-0c4936f540bd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.124809 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8vps\" (UniqueName: \"kubernetes.io/projected/931648f3-732f-4bb1-81d7-6c71f2bf2971-kube-api-access-z8vps\") pod \"redhat-marketplace-96mhl\" (UID: \"931648f3-732f-4bb1-81d7-6c71f2bf2971\") " pod="openshift-marketplace/redhat-marketplace-96mhl" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.124910 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/931648f3-732f-4bb1-81d7-6c71f2bf2971-utilities\") pod \"redhat-marketplace-96mhl\" (UID: \"931648f3-732f-4bb1-81d7-6c71f2bf2971\") " pod="openshift-marketplace/redhat-marketplace-96mhl" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.124982 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b43143bf-1e68-44ef-ac87-0c4936f540bd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b43143bf-1e68-44ef-ac87-0c4936f540bd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.126219 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-plsdd"] Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.135676 4815 patch_prober.go:28] interesting pod/router-default-5444994796-5chwn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 13:22:42 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Feb 25 13:22:42 crc kubenswrapper[4815]: [+]process-running ok Feb 25 13:22:42 crc kubenswrapper[4815]: healthz check failed Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.135725 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5chwn" podUID="e9ae59fe-6c8c-4906-8c36-2b808d0474c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.147989 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.14793951 podStartE2EDuration="2.14793951s" podCreationTimestamp="2026-02-25 13:22:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:42.147000651 +0000 UTC m=+119.948098725" watchObservedRunningTime="2026-02-25 13:22:42.14793951 +0000 UTC m=+119.949037574" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.185576 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4klbc"] Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.188700 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4klbc"] Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.227244 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/931648f3-732f-4bb1-81d7-6c71f2bf2971-catalog-content\") pod \"redhat-marketplace-96mhl\" (UID: \"931648f3-732f-4bb1-81d7-6c71f2bf2971\") " pod="openshift-marketplace/redhat-marketplace-96mhl" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.227312 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b43143bf-1e68-44ef-ac87-0c4936f540bd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b43143bf-1e68-44ef-ac87-0c4936f540bd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.227333 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8vps\" (UniqueName: \"kubernetes.io/projected/931648f3-732f-4bb1-81d7-6c71f2bf2971-kube-api-access-z8vps\") pod \"redhat-marketplace-96mhl\" (UID: \"931648f3-732f-4bb1-81d7-6c71f2bf2971\") " pod="openshift-marketplace/redhat-marketplace-96mhl" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.227357 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/931648f3-732f-4bb1-81d7-6c71f2bf2971-utilities\") pod \"redhat-marketplace-96mhl\" (UID: \"931648f3-732f-4bb1-81d7-6c71f2bf2971\") " pod="openshift-marketplace/redhat-marketplace-96mhl" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.227385 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b43143bf-1e68-44ef-ac87-0c4936f540bd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b43143bf-1e68-44ef-ac87-0c4936f540bd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.228376 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/931648f3-732f-4bb1-81d7-6c71f2bf2971-catalog-content\") pod \"redhat-marketplace-96mhl\" (UID: \"931648f3-732f-4bb1-81d7-6c71f2bf2971\") " pod="openshift-marketplace/redhat-marketplace-96mhl" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.228419 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b43143bf-1e68-44ef-ac87-0c4936f540bd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b43143bf-1e68-44ef-ac87-0c4936f540bd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.229321 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/931648f3-732f-4bb1-81d7-6c71f2bf2971-utilities\") pod \"redhat-marketplace-96mhl\" (UID: \"931648f3-732f-4bb1-81d7-6c71f2bf2971\") " pod="openshift-marketplace/redhat-marketplace-96mhl" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.255564 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d"] Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.258198 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.259125 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b43143bf-1e68-44ef-ac87-0c4936f540bd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b43143bf-1e68-44ef-ac87-0c4936f540bd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.267332 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.273515 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d"] Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.270714 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.270819 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.270861 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.270903 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.270926 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.283180 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8vps\" (UniqueName: \"kubernetes.io/projected/931648f3-732f-4bb1-81d7-6c71f2bf2971-kube-api-access-z8vps\") pod \"redhat-marketplace-96mhl\" (UID: \"931648f3-732f-4bb1-81d7-6c71f2bf2971\") " pod="openshift-marketplace/redhat-marketplace-96mhl" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.321784 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.329741 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7b82\" (UniqueName: \"kubernetes.io/projected/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-kube-api-access-k7b82\") pod \"route-controller-manager-bfc5fbfb9-dgt7d\" (UID: \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\") " pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.329881 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-serving-cert\") pod \"route-controller-manager-bfc5fbfb9-dgt7d\" (UID: \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\") " pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.329904 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-config\") pod \"route-controller-manager-bfc5fbfb9-dgt7d\" (UID: \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\") " pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.329950 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-client-ca\") pod \"route-controller-manager-bfc5fbfb9-dgt7d\" (UID: \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\") " pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.366600 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6zz5"] Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.374823 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z7rhl" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.380495 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.428239 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96mhl" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.431737 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-secret-volume\") pod \"80629dad-7d0a-46b3-bb52-09b75e7dfeb9\" (UID: \"80629dad-7d0a-46b3-bb52-09b75e7dfeb9\") " Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.431865 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w52z\" (UniqueName: \"kubernetes.io/projected/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-kube-api-access-2w52z\") pod \"80629dad-7d0a-46b3-bb52-09b75e7dfeb9\" (UID: \"80629dad-7d0a-46b3-bb52-09b75e7dfeb9\") " Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.431886 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-config-volume\") pod \"80629dad-7d0a-46b3-bb52-09b75e7dfeb9\" (UID: \"80629dad-7d0a-46b3-bb52-09b75e7dfeb9\") " Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.432015 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7b82\" (UniqueName: \"kubernetes.io/projected/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-kube-api-access-k7b82\") pod \"route-controller-manager-bfc5fbfb9-dgt7d\" (UID: \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\") " pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.432084 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-serving-cert\") pod \"route-controller-manager-bfc5fbfb9-dgt7d\" (UID: \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\") " pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.432105 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-config\") pod \"route-controller-manager-bfc5fbfb9-dgt7d\" (UID: \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\") " pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.432124 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-client-ca\") pod \"route-controller-manager-bfc5fbfb9-dgt7d\" (UID: \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\") " pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.435223 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-client-ca\") pod \"route-controller-manager-bfc5fbfb9-dgt7d\" (UID: \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\") " pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.435899 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-config-volume" (OuterVolumeSpecName: "config-volume") pod "80629dad-7d0a-46b3-bb52-09b75e7dfeb9" (UID: "80629dad-7d0a-46b3-bb52-09b75e7dfeb9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.437633 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-kube-api-access-2w52z" (OuterVolumeSpecName: "kube-api-access-2w52z") pod "80629dad-7d0a-46b3-bb52-09b75e7dfeb9" (UID: "80629dad-7d0a-46b3-bb52-09b75e7dfeb9"). InnerVolumeSpecName "kube-api-access-2w52z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.438037 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-config\") pod \"route-controller-manager-bfc5fbfb9-dgt7d\" (UID: \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\") " pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.441770 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "80629dad-7d0a-46b3-bb52-09b75e7dfeb9" (UID: "80629dad-7d0a-46b3-bb52-09b75e7dfeb9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.442371 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-serving-cert\") pod \"route-controller-manager-bfc5fbfb9-dgt7d\" (UID: \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\") " pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.456085 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7b82\" (UniqueName: \"kubernetes.io/projected/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-kube-api-access-k7b82\") pod \"route-controller-manager-bfc5fbfb9-dgt7d\" (UID: \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\") " pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.533280 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w52z\" (UniqueName: \"kubernetes.io/projected/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-kube-api-access-2w52z\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.533307 4815 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.533316 4815 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80629dad-7d0a-46b3-bb52-09b75e7dfeb9-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.555954 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f"] Feb 25 13:22:42 crc kubenswrapper[4815]: E0225 13:22:42.556166 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80629dad-7d0a-46b3-bb52-09b75e7dfeb9" containerName="collect-profiles" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.556178 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="80629dad-7d0a-46b3-bb52-09b75e7dfeb9" containerName="collect-profiles" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.556261 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="80629dad-7d0a-46b3-bb52-09b75e7dfeb9" containerName="collect-profiles" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.556618 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.562964 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.563125 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.563204 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.563295 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.563426 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.568417 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.568727 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.574189 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f"] Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.594250 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.611923 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.634365 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdt4c\" (UniqueName: \"kubernetes.io/projected/5577840d-a3d2-4606-8e4f-4f74401a9b2f-kube-api-access-tdt4c\") pod \"controller-manager-c7ccfc5c7-kfv8f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.634706 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-config\") pod \"controller-manager-c7ccfc5c7-kfv8f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.634769 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5577840d-a3d2-4606-8e4f-4f74401a9b2f-serving-cert\") pod \"controller-manager-c7ccfc5c7-kfv8f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.634793 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-proxy-ca-bundles\") pod \"controller-manager-c7ccfc5c7-kfv8f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.634845 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-client-ca\") pod \"controller-manager-c7ccfc5c7-kfv8f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.685590 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9pjrm"] Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.688195 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9pjrm" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.690126 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.692460 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9pjrm"] Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.740497 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-config\") pod \"controller-manager-c7ccfc5c7-kfv8f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.740559 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-catalog-content\") pod \"redhat-operators-9pjrm\" (UID: \"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4\") " pod="openshift-marketplace/redhat-operators-9pjrm" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.740605 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz6fn\" (UniqueName: \"kubernetes.io/projected/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-kube-api-access-cz6fn\") pod \"redhat-operators-9pjrm\" (UID: \"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4\") " pod="openshift-marketplace/redhat-operators-9pjrm" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.740637 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-utilities\") pod \"redhat-operators-9pjrm\" (UID: \"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4\") " pod="openshift-marketplace/redhat-operators-9pjrm" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.740664 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5577840d-a3d2-4606-8e4f-4f74401a9b2f-serving-cert\") pod \"controller-manager-c7ccfc5c7-kfv8f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.740694 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-proxy-ca-bundles\") pod \"controller-manager-c7ccfc5c7-kfv8f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.740748 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-client-ca\") pod \"controller-manager-c7ccfc5c7-kfv8f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.740781 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdt4c\" (UniqueName: \"kubernetes.io/projected/5577840d-a3d2-4606-8e4f-4f74401a9b2f-kube-api-access-tdt4c\") pod \"controller-manager-c7ccfc5c7-kfv8f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.742432 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-config\") pod \"controller-manager-c7ccfc5c7-kfv8f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.742693 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-proxy-ca-bundles\") pod \"controller-manager-c7ccfc5c7-kfv8f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.743127 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-client-ca\") pod \"controller-manager-c7ccfc5c7-kfv8f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.743873 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-96mhl"] Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.748041 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5577840d-a3d2-4606-8e4f-4f74401a9b2f-serving-cert\") pod \"controller-manager-c7ccfc5c7-kfv8f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.770398 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdt4c\" (UniqueName: \"kubernetes.io/projected/5577840d-a3d2-4606-8e4f-4f74401a9b2f-kube-api-access-tdt4c\") pod \"controller-manager-c7ccfc5c7-kfv8f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.841629 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-catalog-content\") pod \"redhat-operators-9pjrm\" (UID: \"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4\") " pod="openshift-marketplace/redhat-operators-9pjrm" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.841687 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz6fn\" (UniqueName: \"kubernetes.io/projected/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-kube-api-access-cz6fn\") pod \"redhat-operators-9pjrm\" (UID: \"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4\") " pod="openshift-marketplace/redhat-operators-9pjrm" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.841712 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-utilities\") pod \"redhat-operators-9pjrm\" (UID: \"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4\") " pod="openshift-marketplace/redhat-operators-9pjrm" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.842335 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-utilities\") pod \"redhat-operators-9pjrm\" (UID: \"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4\") " pod="openshift-marketplace/redhat-operators-9pjrm" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.842606 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-catalog-content\") pod \"redhat-operators-9pjrm\" (UID: \"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4\") " pod="openshift-marketplace/redhat-operators-9pjrm" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.873236 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz6fn\" (UniqueName: \"kubernetes.io/projected/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-kube-api-access-cz6fn\") pod \"redhat-operators-9pjrm\" (UID: \"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4\") " pod="openshift-marketplace/redhat-operators-9pjrm" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.874196 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d"] Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.901660 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.910792 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.948002 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60971790-c5dc-4884-ae44-e91d035cbe31" path="/var/lib/kubelet/pods/60971790-c5dc-4884-ae44-e91d035cbe31/volumes" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.948762 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72d966ad-816d-4483-b9e6-44edb10f9d75" path="/var/lib/kubelet/pods/72d966ad-816d-4483-b9e6-44edb10f9d75/volumes" Feb 25 13:22:42 crc kubenswrapper[4815]: I0225 13:22:42.949399 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.011199 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.021156 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9pjrm" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.026295 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96mhl" event={"ID":"931648f3-732f-4bb1-81d7-6c71f2bf2971","Type":"ContainerStarted","Data":"1b773db05120718fa34285eba8e410e83b89ef80f85f20c8099eee116b344734"} Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.029128 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" event={"ID":"072fbe81-88e1-4755-bb59-5ee95d72760f","Type":"ContainerStarted","Data":"4a0b4fa3f4b8e7df475878429dc453ee25d3a6301460b194191752957741cd60"} Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.029707 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.066772 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.066848 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv" event={"ID":"80629dad-7d0a-46b3-bb52-09b75e7dfeb9","Type":"ContainerDied","Data":"28a7fa6c0beba75fc11d190790b4525724e33b78a6d798bacdab35592dc0971c"} Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.067641 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28a7fa6c0beba75fc11d190790b4525724e33b78a6d798bacdab35592dc0971c" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.083933 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hbtkx"] Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.085751 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hbtkx" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.099115 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b43143bf-1e68-44ef-ac87-0c4936f540bd","Type":"ContainerStarted","Data":"173fc053534645077ce27ecfa6934de00d75063cae5d71c11f82e873d1ae5739"} Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.099878 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hbtkx"] Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.115719 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" podStartSLOduration=68.115696267 podStartE2EDuration="1m8.115696267s" podCreationTimestamp="2026-02-25 13:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:43.113059947 +0000 UTC m=+120.914158011" watchObservedRunningTime="2026-02-25 13:22:43.115696267 +0000 UTC m=+120.916794331" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.119436 4815 generic.go:334] "Generic (PLEG): container finished" podID="6d00f8b4-1bf6-4a47-9beb-2226f26d695b" containerID="d8baf2b15896c3c7a59be66e98f35e71c42786196fb673aa0de2dc1894be3465" exitCode=0 Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.119617 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6zz5" event={"ID":"6d00f8b4-1bf6-4a47-9beb-2226f26d695b","Type":"ContainerDied","Data":"d8baf2b15896c3c7a59be66e98f35e71c42786196fb673aa0de2dc1894be3465"} Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.119682 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6zz5" event={"ID":"6d00f8b4-1bf6-4a47-9beb-2226f26d695b","Type":"ContainerStarted","Data":"0a5ffda5565243ea6d11416bdfcc0373a0a1457db88d02797cd1097a8be04ab7"} Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.130987 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.132355 4815 generic.go:334] "Generic (PLEG): container finished" podID="c846b62d-3716-43ac-8f8d-b68c593f3b49" containerID="d77bc3984367ebd298a4511fccf39ef8b8d4ddbc459782d2dddcd1a28f773bc5" exitCode=0 Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.132449 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c846b62d-3716-43ac-8f8d-b68c593f3b49","Type":"ContainerDied","Data":"d77bc3984367ebd298a4511fccf39ef8b8d4ddbc459782d2dddcd1a28f773bc5"} Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.134464 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" event={"ID":"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8","Type":"ContainerStarted","Data":"73ed31567e34138350e582d640ccb5a5d6ec79dffd8ecc62a2f1dcdc51e214bc"} Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.138435 4815 patch_prober.go:28] interesting pod/router-default-5444994796-5chwn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 13:22:43 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Feb 25 13:22:43 crc kubenswrapper[4815]: [+]process-running ok Feb 25 13:22:43 crc kubenswrapper[4815]: healthz check failed Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.138480 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5chwn" podUID="e9ae59fe-6c8c-4906-8c36-2b808d0474c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.222550 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-8jnnl container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" start-of-body= Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.222849 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-8jnnl" podUID="78d30b59-0a66-44e5-b0b9-a2864e4d4552" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.222750 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-8jnnl container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" start-of-body= Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.222953 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8jnnl" podUID="78d30b59-0a66-44e5-b0b9-a2864e4d4552" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.253031 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xkdr\" (UniqueName: \"kubernetes.io/projected/72ffb200-965d-41f3-9900-efba5e409b8d-kube-api-access-5xkdr\") pod \"redhat-operators-hbtkx\" (UID: \"72ffb200-965d-41f3-9900-efba5e409b8d\") " pod="openshift-marketplace/redhat-operators-hbtkx" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.253152 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72ffb200-965d-41f3-9900-efba5e409b8d-utilities\") pod \"redhat-operators-hbtkx\" (UID: \"72ffb200-965d-41f3-9900-efba5e409b8d\") " pod="openshift-marketplace/redhat-operators-hbtkx" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.253189 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72ffb200-965d-41f3-9900-efba5e409b8d-catalog-content\") pod \"redhat-operators-hbtkx\" (UID: \"72ffb200-965d-41f3-9900-efba5e409b8d\") " pod="openshift-marketplace/redhat-operators-hbtkx" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.353940 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xkdr\" (UniqueName: \"kubernetes.io/projected/72ffb200-965d-41f3-9900-efba5e409b8d-kube-api-access-5xkdr\") pod \"redhat-operators-hbtkx\" (UID: \"72ffb200-965d-41f3-9900-efba5e409b8d\") " pod="openshift-marketplace/redhat-operators-hbtkx" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.354044 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72ffb200-965d-41f3-9900-efba5e409b8d-utilities\") pod \"redhat-operators-hbtkx\" (UID: \"72ffb200-965d-41f3-9900-efba5e409b8d\") " pod="openshift-marketplace/redhat-operators-hbtkx" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.354078 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72ffb200-965d-41f3-9900-efba5e409b8d-catalog-content\") pod \"redhat-operators-hbtkx\" (UID: \"72ffb200-965d-41f3-9900-efba5e409b8d\") " pod="openshift-marketplace/redhat-operators-hbtkx" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.355035 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72ffb200-965d-41f3-9900-efba5e409b8d-catalog-content\") pod \"redhat-operators-hbtkx\" (UID: \"72ffb200-965d-41f3-9900-efba5e409b8d\") " pod="openshift-marketplace/redhat-operators-hbtkx" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.355214 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72ffb200-965d-41f3-9900-efba5e409b8d-utilities\") pod \"redhat-operators-hbtkx\" (UID: \"72ffb200-965d-41f3-9900-efba5e409b8d\") " pod="openshift-marketplace/redhat-operators-hbtkx" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.372554 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f"] Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.374576 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xkdr\" (UniqueName: \"kubernetes.io/projected/72ffb200-965d-41f3-9900-efba5e409b8d-kube-api-access-5xkdr\") pod \"redhat-operators-hbtkx\" (UID: \"72ffb200-965d-41f3-9900-efba5e409b8d\") " pod="openshift-marketplace/redhat-operators-hbtkx" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.387872 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.389379 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.411682 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9pjrm"] Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.414800 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:43 crc kubenswrapper[4815]: W0225 13:22:43.421875 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbdf40b79_9414_4062_8dfd_3f0f71fcd1e4.slice/crio-bdae89adda41be4ab717561875f8ddbcd505a0da76baa92e429fd3508d056cd4 WatchSource:0}: Error finding container bdae89adda41be4ab717561875f8ddbcd505a0da76baa92e429fd3508d056cd4: Status 404 returned error can't find the container with id bdae89adda41be4ab717561875f8ddbcd505a0da76baa92e429fd3508d056cd4 Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.432696 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hbtkx" Feb 25 13:22:43 crc kubenswrapper[4815]: E0225 13:22:43.738456 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 13:22:43 crc kubenswrapper[4815]: E0225 13:22:43.741855 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 13:22:43 crc kubenswrapper[4815]: E0225 13:22:43.743876 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 13:22:43 crc kubenswrapper[4815]: E0225 13:22:43.743914 4815 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" podUID="65dfccf6-aa91-41fa-95d6-fd4778065408" containerName="kube-multus-additional-cni-plugins" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.884007 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.884364 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.884668 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.896910 4815 patch_prober.go:28] interesting pod/console-f9d7485db-mhgxd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.896963 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-mhgxd" podUID="a81edad3-179a-4786-9ae9-e1a8f0a4d01a" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.898971 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qbnmz" Feb 25 13:22:43 crc kubenswrapper[4815]: I0225 13:22:43.916251 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hbtkx"] Feb 25 13:22:43 crc kubenswrapper[4815]: W0225 13:22:43.953127 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72ffb200_965d_41f3_9900_efba5e409b8d.slice/crio-b581c3a933ce725d120845df267555a60feddd8f3e7317026c06e7850d765c14 WatchSource:0}: Error finding container b581c3a933ce725d120845df267555a60feddd8f3e7317026c06e7850d765c14: Status 404 returned error can't find the container with id b581c3a933ce725d120845df267555a60feddd8f3e7317026c06e7850d765c14 Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.134265 4815 patch_prober.go:28] interesting pod/router-default-5444994796-5chwn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 13:22:44 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Feb 25 13:22:44 crc kubenswrapper[4815]: [+]process-running ok Feb 25 13:22:44 crc kubenswrapper[4815]: healthz check failed Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.134306 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5chwn" podUID="e9ae59fe-6c8c-4906-8c36-2b808d0474c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.145914 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" event={"ID":"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8","Type":"ContainerStarted","Data":"8f12b2e55ea7cc8e5842c756e5112c363ec72a99595bbe8616262b8642c38a4e"} Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.146546 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.152971 4815 generic.go:334] "Generic (PLEG): container finished" podID="bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" containerID="5b2a897650c0e7a04106ae14a057313aef7644492d09282509472eee678d235c" exitCode=0 Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.154054 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9pjrm" event={"ID":"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4","Type":"ContainerDied","Data":"5b2a897650c0e7a04106ae14a057313aef7644492d09282509472eee678d235c"} Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.154288 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9pjrm" event={"ID":"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4","Type":"ContainerStarted","Data":"bdae89adda41be4ab717561875f8ddbcd505a0da76baa92e429fd3508d056cd4"} Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.158324 4815 generic.go:334] "Generic (PLEG): container finished" podID="b43143bf-1e68-44ef-ac87-0c4936f540bd" containerID="5e23840fc195155a6a98c3d9960415a4c49a92863cd0b804cada3866b39904f7" exitCode=0 Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.158385 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b43143bf-1e68-44ef-ac87-0c4936f540bd","Type":"ContainerDied","Data":"5e23840fc195155a6a98c3d9960415a4c49a92863cd0b804cada3866b39904f7"} Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.162443 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.163783 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" podStartSLOduration=2.163773353 podStartE2EDuration="2.163773353s" podCreationTimestamp="2026-02-25 13:22:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:44.162747551 +0000 UTC m=+121.963845605" watchObservedRunningTime="2026-02-25 13:22:44.163773353 +0000 UTC m=+121.964871407" Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.173016 4815 generic.go:334] "Generic (PLEG): container finished" podID="931648f3-732f-4bb1-81d7-6c71f2bf2971" containerID="cecc9c8ae4ef2e44d1df26f3cb5804514de160ff0579bfb126efdf7c2252f1a9" exitCode=0 Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.173141 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96mhl" event={"ID":"931648f3-732f-4bb1-81d7-6c71f2bf2971","Type":"ContainerDied","Data":"cecc9c8ae4ef2e44d1df26f3cb5804514de160ff0579bfb126efdf7c2252f1a9"} Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.177449 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbtkx" event={"ID":"72ffb200-965d-41f3-9900-efba5e409b8d","Type":"ContainerStarted","Data":"b581c3a933ce725d120845df267555a60feddd8f3e7317026c06e7850d765c14"} Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.189450 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" event={"ID":"5577840d-a3d2-4606-8e4f-4f74401a9b2f","Type":"ContainerStarted","Data":"b324e157d608d68382517616a269450458833a82bfa6bbbeea58aa381866aecf"} Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.189752 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" event={"ID":"5577840d-a3d2-4606-8e4f-4f74401a9b2f","Type":"ContainerStarted","Data":"d46265344c6eda0effb02957a4a04ebbc92361d38f83913d4f3bb1d8b754fcae"} Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.190277 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.205236 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.208153 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-hckbl" Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.250105 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" podStartSLOduration=4.250080613 podStartE2EDuration="4.250080613s" podCreationTimestamp="2026-02-25 13:22:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:44.244682429 +0000 UTC m=+122.045780503" watchObservedRunningTime="2026-02-25 13:22:44.250080613 +0000 UTC m=+122.051178667" Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.606884 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.680367 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c846b62d-3716-43ac-8f8d-b68c593f3b49-kube-api-access\") pod \"c846b62d-3716-43ac-8f8d-b68c593f3b49\" (UID: \"c846b62d-3716-43ac-8f8d-b68c593f3b49\") " Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.680478 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c846b62d-3716-43ac-8f8d-b68c593f3b49-kubelet-dir\") pod \"c846b62d-3716-43ac-8f8d-b68c593f3b49\" (UID: \"c846b62d-3716-43ac-8f8d-b68c593f3b49\") " Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.680730 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c846b62d-3716-43ac-8f8d-b68c593f3b49-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c846b62d-3716-43ac-8f8d-b68c593f3b49" (UID: "c846b62d-3716-43ac-8f8d-b68c593f3b49"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.687563 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c846b62d-3716-43ac-8f8d-b68c593f3b49-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c846b62d-3716-43ac-8f8d-b68c593f3b49" (UID: "c846b62d-3716-43ac-8f8d-b68c593f3b49"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.782690 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c846b62d-3716-43ac-8f8d-b68c593f3b49-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:44 crc kubenswrapper[4815]: I0225 13:22:44.783033 4815 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c846b62d-3716-43ac-8f8d-b68c593f3b49-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:45 crc kubenswrapper[4815]: I0225 13:22:45.145036 4815 patch_prober.go:28] interesting pod/router-default-5444994796-5chwn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 13:22:45 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Feb 25 13:22:45 crc kubenswrapper[4815]: [+]process-running ok Feb 25 13:22:45 crc kubenswrapper[4815]: healthz check failed Feb 25 13:22:45 crc kubenswrapper[4815]: I0225 13:22:45.145102 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5chwn" podUID="e9ae59fe-6c8c-4906-8c36-2b808d0474c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 13:22:45 crc kubenswrapper[4815]: I0225 13:22:45.208646 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c846b62d-3716-43ac-8f8d-b68c593f3b49","Type":"ContainerDied","Data":"dd1540c82e4c070ffe35b6e62f4def0f166b15f46cd1b7c78221300c51d0cc55"} Feb 25 13:22:45 crc kubenswrapper[4815]: I0225 13:22:45.208690 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd1540c82e4c070ffe35b6e62f4def0f166b15f46cd1b7c78221300c51d0cc55" Feb 25 13:22:45 crc kubenswrapper[4815]: I0225 13:22:45.208769 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 13:22:45 crc kubenswrapper[4815]: I0225 13:22:45.212347 4815 generic.go:334] "Generic (PLEG): container finished" podID="72ffb200-965d-41f3-9900-efba5e409b8d" containerID="d8aa12ff096ee354b371f67ee8a4bb21f9987ee44fa58067b111355b0c744e06" exitCode=0 Feb 25 13:22:45 crc kubenswrapper[4815]: I0225 13:22:45.212490 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbtkx" event={"ID":"72ffb200-965d-41f3-9900-efba5e409b8d","Type":"ContainerDied","Data":"d8aa12ff096ee354b371f67ee8a4bb21f9987ee44fa58067b111355b0c744e06"} Feb 25 13:22:45 crc kubenswrapper[4815]: I0225 13:22:45.572095 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 13:22:45 crc kubenswrapper[4815]: I0225 13:22:45.694471 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b43143bf-1e68-44ef-ac87-0c4936f540bd-kubelet-dir\") pod \"b43143bf-1e68-44ef-ac87-0c4936f540bd\" (UID: \"b43143bf-1e68-44ef-ac87-0c4936f540bd\") " Feb 25 13:22:45 crc kubenswrapper[4815]: I0225 13:22:45.694564 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b43143bf-1e68-44ef-ac87-0c4936f540bd-kube-api-access\") pod \"b43143bf-1e68-44ef-ac87-0c4936f540bd\" (UID: \"b43143bf-1e68-44ef-ac87-0c4936f540bd\") " Feb 25 13:22:45 crc kubenswrapper[4815]: I0225 13:22:45.694985 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b43143bf-1e68-44ef-ac87-0c4936f540bd-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b43143bf-1e68-44ef-ac87-0c4936f540bd" (UID: "b43143bf-1e68-44ef-ac87-0c4936f540bd"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:22:45 crc kubenswrapper[4815]: I0225 13:22:45.718200 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b43143bf-1e68-44ef-ac87-0c4936f540bd-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b43143bf-1e68-44ef-ac87-0c4936f540bd" (UID: "b43143bf-1e68-44ef-ac87-0c4936f540bd"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:22:45 crc kubenswrapper[4815]: I0225 13:22:45.796429 4815 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b43143bf-1e68-44ef-ac87-0c4936f540bd-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:45 crc kubenswrapper[4815]: I0225 13:22:45.796478 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b43143bf-1e68-44ef-ac87-0c4936f540bd-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 25 13:22:45 crc kubenswrapper[4815]: I0225 13:22:45.861068 4815 ???:1] "http: TLS handshake error from 192.168.126.11:59484: no serving certificate available for the kubelet" Feb 25 13:22:46 crc kubenswrapper[4815]: I0225 13:22:46.045810 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-kw6xm" Feb 25 13:22:46 crc kubenswrapper[4815]: I0225 13:22:46.132904 4815 patch_prober.go:28] interesting pod/router-default-5444994796-5chwn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 13:22:46 crc kubenswrapper[4815]: [+]has-synced ok Feb 25 13:22:46 crc kubenswrapper[4815]: [+]process-running ok Feb 25 13:22:46 crc kubenswrapper[4815]: healthz check failed Feb 25 13:22:46 crc kubenswrapper[4815]: I0225 13:22:46.133137 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5chwn" podUID="e9ae59fe-6c8c-4906-8c36-2b808d0474c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 13:22:46 crc kubenswrapper[4815]: I0225 13:22:46.240785 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b43143bf-1e68-44ef-ac87-0c4936f540bd","Type":"ContainerDied","Data":"173fc053534645077ce27ecfa6934de00d75063cae5d71c11f82e873d1ae5739"} Feb 25 13:22:46 crc kubenswrapper[4815]: I0225 13:22:46.240832 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="173fc053534645077ce27ecfa6934de00d75063cae5d71c11f82e873d1ae5739" Feb 25 13:22:46 crc kubenswrapper[4815]: I0225 13:22:46.240892 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 13:22:47 crc kubenswrapper[4815]: I0225 13:22:47.025919 4815 ???:1] "http: TLS handshake error from 192.168.126.11:59496: no serving certificate available for the kubelet" Feb 25 13:22:47 crc kubenswrapper[4815]: I0225 13:22:47.132318 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:47 crc kubenswrapper[4815]: I0225 13:22:47.134629 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-5chwn" Feb 25 13:22:49 crc kubenswrapper[4815]: I0225 13:22:49.920406 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:22:53 crc kubenswrapper[4815]: I0225 13:22:53.228344 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-8jnnl" Feb 25 13:22:53 crc kubenswrapper[4815]: E0225 13:22:53.736281 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 13:22:53 crc kubenswrapper[4815]: E0225 13:22:53.737859 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 13:22:53 crc kubenswrapper[4815]: E0225 13:22:53.739161 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 13:22:53 crc kubenswrapper[4815]: E0225 13:22:53.739195 4815 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" podUID="65dfccf6-aa91-41fa-95d6-fd4778065408" containerName="kube-multus-additional-cni-plugins" Feb 25 13:22:53 crc kubenswrapper[4815]: I0225 13:22:53.888466 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:53 crc kubenswrapper[4815]: I0225 13:22:53.895461 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:22:56 crc kubenswrapper[4815]: I0225 13:22:56.123623 4815 ???:1] "http: TLS handshake error from 192.168.126.11:32894: no serving certificate available for the kubelet" Feb 25 13:22:58 crc kubenswrapper[4815]: I0225 13:22:58.959752 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 25 13:22:59 crc kubenswrapper[4815]: I0225 13:22:59.398552 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f"] Feb 25 13:22:59 crc kubenswrapper[4815]: I0225 13:22:59.398886 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" podUID="5577840d-a3d2-4606-8e4f-4f74401a9b2f" containerName="controller-manager" containerID="cri-o://b324e157d608d68382517616a269450458833a82bfa6bbbeea58aa381866aecf" gracePeriod=30 Feb 25 13:22:59 crc kubenswrapper[4815]: I0225 13:22:59.420965 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d"] Feb 25 13:22:59 crc kubenswrapper[4815]: I0225 13:22:59.421176 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" podUID="482c5aac-8d62-4d76-8b31-ac99c1cfc4e8" containerName="route-controller-manager" containerID="cri-o://8f12b2e55ea7cc8e5842c756e5112c363ec72a99595bbe8616262b8642c38a4e" gracePeriod=30 Feb 25 13:22:59 crc kubenswrapper[4815]: I0225 13:22:59.426255 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=1.426238968 podStartE2EDuration="1.426238968s" podCreationTimestamp="2026-02-25 13:22:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:22:59.424676614 +0000 UTC m=+137.225774678" watchObservedRunningTime="2026-02-25 13:22:59.426238968 +0000 UTC m=+137.227337022" Feb 25 13:23:01 crc kubenswrapper[4815]: I0225 13:23:01.312880 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:23:01 crc kubenswrapper[4815]: I0225 13:23:01.366060 4815 generic.go:334] "Generic (PLEG): container finished" podID="482c5aac-8d62-4d76-8b31-ac99c1cfc4e8" containerID="8f12b2e55ea7cc8e5842c756e5112c363ec72a99595bbe8616262b8642c38a4e" exitCode=0 Feb 25 13:23:01 crc kubenswrapper[4815]: I0225 13:23:01.366140 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" event={"ID":"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8","Type":"ContainerDied","Data":"8f12b2e55ea7cc8e5842c756e5112c363ec72a99595bbe8616262b8642c38a4e"} Feb 25 13:23:01 crc kubenswrapper[4815]: I0225 13:23:01.368759 4815 generic.go:334] "Generic (PLEG): container finished" podID="5577840d-a3d2-4606-8e4f-4f74401a9b2f" containerID="b324e157d608d68382517616a269450458833a82bfa6bbbeea58aa381866aecf" exitCode=0 Feb 25 13:23:01 crc kubenswrapper[4815]: I0225 13:23:01.368830 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" event={"ID":"5577840d-a3d2-4606-8e4f-4f74401a9b2f","Type":"ContainerDied","Data":"b324e157d608d68382517616a269450458833a82bfa6bbbeea58aa381866aecf"} Feb 25 13:23:02 crc kubenswrapper[4815]: I0225 13:23:02.597470 4815 patch_prober.go:28] interesting pod/route-controller-manager-bfc5fbfb9-dgt7d container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" start-of-body= Feb 25 13:23:02 crc kubenswrapper[4815]: I0225 13:23:02.597634 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" podUID="482c5aac-8d62-4d76-8b31-ac99c1cfc4e8" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" Feb 25 13:23:02 crc kubenswrapper[4815]: I0225 13:23:02.912186 4815 patch_prober.go:28] interesting pod/controller-manager-c7ccfc5c7-kfv8f container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.54:8443/healthz\": dial tcp 10.217.0.54:8443: connect: connection refused" start-of-body= Feb 25 13:23:02 crc kubenswrapper[4815]: I0225 13:23:02.912258 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" podUID="5577840d-a3d2-4606-8e4f-4f74401a9b2f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.54:8443/healthz\": dial tcp 10.217.0.54:8443: connect: connection refused" Feb 25 13:23:03 crc kubenswrapper[4815]: E0225 13:23:03.736347 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 13:23:03 crc kubenswrapper[4815]: E0225 13:23:03.737610 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 13:23:03 crc kubenswrapper[4815]: E0225 13:23:03.738887 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 13:23:03 crc kubenswrapper[4815]: E0225 13:23:03.738943 4815 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" podUID="65dfccf6-aa91-41fa-95d6-fd4778065408" containerName="kube-multus-additional-cni-plugins" Feb 25 13:23:09 crc kubenswrapper[4815]: E0225 13:23:09.674906 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Feb 25 13:23:09 crc kubenswrapper[4815]: E0225 13:23:09.675479 4815 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 25 13:23:09 crc kubenswrapper[4815]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Feb 25 13:23:09 crc kubenswrapper[4815]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7k82t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29533762-lkzv9_openshift-infra(6403af09-14ef-49fc-858f-06c1bbadb88b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Feb 25 13:23:09 crc kubenswrapper[4815]: > logger="UnhandledError" Feb 25 13:23:09 crc kubenswrapper[4815]: E0225 13:23:09.676717 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29533762-lkzv9" podUID="6403af09-14ef-49fc-858f-06c1bbadb88b" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.698218 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.702530 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.736251 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-client-ca\") pod \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.736316 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-proxy-ca-bundles\") pod \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.736343 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-serving-cert\") pod \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\" (UID: \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\") " Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.736366 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-config\") pod \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\" (UID: \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\") " Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.736418 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdt4c\" (UniqueName: \"kubernetes.io/projected/5577840d-a3d2-4606-8e4f-4f74401a9b2f-kube-api-access-tdt4c\") pod \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.736468 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-config\") pod \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.736489 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-client-ca\") pod \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\" (UID: \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\") " Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.736538 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5577840d-a3d2-4606-8e4f-4f74401a9b2f-serving-cert\") pod \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\" (UID: \"5577840d-a3d2-4606-8e4f-4f74401a9b2f\") " Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.736708 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7b82\" (UniqueName: \"kubernetes.io/projected/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-kube-api-access-k7b82\") pod \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\" (UID: \"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8\") " Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.737093 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-client-ca" (OuterVolumeSpecName: "client-ca") pod "5577840d-a3d2-4606-8e4f-4f74401a9b2f" (UID: "5577840d-a3d2-4606-8e4f-4f74401a9b2f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.738795 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-client-ca" (OuterVolumeSpecName: "client-ca") pod "482c5aac-8d62-4d76-8b31-ac99c1cfc4e8" (UID: "482c5aac-8d62-4d76-8b31-ac99c1cfc4e8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.739440 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52"] Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.739882 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-config" (OuterVolumeSpecName: "config") pod "482c5aac-8d62-4d76-8b31-ac99c1cfc4e8" (UID: "482c5aac-8d62-4d76-8b31-ac99c1cfc4e8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.739894 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-config" (OuterVolumeSpecName: "config") pod "5577840d-a3d2-4606-8e4f-4f74401a9b2f" (UID: "5577840d-a3d2-4606-8e4f-4f74401a9b2f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:23:09 crc kubenswrapper[4815]: E0225 13:23:09.739933 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5577840d-a3d2-4606-8e4f-4f74401a9b2f" containerName="controller-manager" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.739953 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="5577840d-a3d2-4606-8e4f-4f74401a9b2f" containerName="controller-manager" Feb 25 13:23:09 crc kubenswrapper[4815]: E0225 13:23:09.739974 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b43143bf-1e68-44ef-ac87-0c4936f540bd" containerName="pruner" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.739981 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b43143bf-1e68-44ef-ac87-0c4936f540bd" containerName="pruner" Feb 25 13:23:09 crc kubenswrapper[4815]: E0225 13:23:09.739995 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="482c5aac-8d62-4d76-8b31-ac99c1cfc4e8" containerName="route-controller-manager" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.740004 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="482c5aac-8d62-4d76-8b31-ac99c1cfc4e8" containerName="route-controller-manager" Feb 25 13:23:09 crc kubenswrapper[4815]: E0225 13:23:09.740017 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c846b62d-3716-43ac-8f8d-b68c593f3b49" containerName="pruner" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.740024 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="c846b62d-3716-43ac-8f8d-b68c593f3b49" containerName="pruner" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.740008 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "5577840d-a3d2-4606-8e4f-4f74401a9b2f" (UID: "5577840d-a3d2-4606-8e4f-4f74401a9b2f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.740222 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="5577840d-a3d2-4606-8e4f-4f74401a9b2f" containerName="controller-manager" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.740236 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="482c5aac-8d62-4d76-8b31-ac99c1cfc4e8" containerName="route-controller-manager" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.740252 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="c846b62d-3716-43ac-8f8d-b68c593f3b49" containerName="pruner" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.740262 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b43143bf-1e68-44ef-ac87-0c4936f540bd" containerName="pruner" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.741854 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.748630 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5577840d-a3d2-4606-8e4f-4f74401a9b2f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5577840d-a3d2-4606-8e4f-4f74401a9b2f" (UID: "5577840d-a3d2-4606-8e4f-4f74401a9b2f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.748739 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "482c5aac-8d62-4d76-8b31-ac99c1cfc4e8" (UID: "482c5aac-8d62-4d76-8b31-ac99c1cfc4e8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.749061 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5577840d-a3d2-4606-8e4f-4f74401a9b2f-kube-api-access-tdt4c" (OuterVolumeSpecName: "kube-api-access-tdt4c") pod "5577840d-a3d2-4606-8e4f-4f74401a9b2f" (UID: "5577840d-a3d2-4606-8e4f-4f74401a9b2f"). InnerVolumeSpecName "kube-api-access-tdt4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.753681 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-kube-api-access-k7b82" (OuterVolumeSpecName: "kube-api-access-k7b82") pod "482c5aac-8d62-4d76-8b31-ac99c1cfc4e8" (UID: "482c5aac-8d62-4d76-8b31-ac99c1cfc4e8"). InnerVolumeSpecName "kube-api-access-k7b82". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.755589 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52"] Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.837828 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d1f4a507-b974-41a1-bdd8-c32458306d9d-client-ca\") pod \"route-controller-manager-5df4b7c597-5xl52\" (UID: \"d1f4a507-b974-41a1-bdd8-c32458306d9d\") " pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.837888 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1f4a507-b974-41a1-bdd8-c32458306d9d-config\") pod \"route-controller-manager-5df4b7c597-5xl52\" (UID: \"d1f4a507-b974-41a1-bdd8-c32458306d9d\") " pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.837987 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l9vk\" (UniqueName: \"kubernetes.io/projected/d1f4a507-b974-41a1-bdd8-c32458306d9d-kube-api-access-9l9vk\") pod \"route-controller-manager-5df4b7c597-5xl52\" (UID: \"d1f4a507-b974-41a1-bdd8-c32458306d9d\") " pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.838107 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1f4a507-b974-41a1-bdd8-c32458306d9d-serving-cert\") pod \"route-controller-manager-5df4b7c597-5xl52\" (UID: \"d1f4a507-b974-41a1-bdd8-c32458306d9d\") " pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.838195 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.838207 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdt4c\" (UniqueName: \"kubernetes.io/projected/5577840d-a3d2-4606-8e4f-4f74401a9b2f-kube-api-access-tdt4c\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.838219 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.838227 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.838237 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5577840d-a3d2-4606-8e4f-4f74401a9b2f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.838246 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7b82\" (UniqueName: \"kubernetes.io/projected/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-kube-api-access-k7b82\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.838254 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.838262 4815 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5577840d-a3d2-4606-8e4f-4f74401a9b2f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.838271 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.939620 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d1f4a507-b974-41a1-bdd8-c32458306d9d-client-ca\") pod \"route-controller-manager-5df4b7c597-5xl52\" (UID: \"d1f4a507-b974-41a1-bdd8-c32458306d9d\") " pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.939710 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1f4a507-b974-41a1-bdd8-c32458306d9d-config\") pod \"route-controller-manager-5df4b7c597-5xl52\" (UID: \"d1f4a507-b974-41a1-bdd8-c32458306d9d\") " pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.939759 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l9vk\" (UniqueName: \"kubernetes.io/projected/d1f4a507-b974-41a1-bdd8-c32458306d9d-kube-api-access-9l9vk\") pod \"route-controller-manager-5df4b7c597-5xl52\" (UID: \"d1f4a507-b974-41a1-bdd8-c32458306d9d\") " pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.939788 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1f4a507-b974-41a1-bdd8-c32458306d9d-serving-cert\") pod \"route-controller-manager-5df4b7c597-5xl52\" (UID: \"d1f4a507-b974-41a1-bdd8-c32458306d9d\") " pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.940577 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d1f4a507-b974-41a1-bdd8-c32458306d9d-client-ca\") pod \"route-controller-manager-5df4b7c597-5xl52\" (UID: \"d1f4a507-b974-41a1-bdd8-c32458306d9d\") " pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.943112 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1f4a507-b974-41a1-bdd8-c32458306d9d-config\") pod \"route-controller-manager-5df4b7c597-5xl52\" (UID: \"d1f4a507-b974-41a1-bdd8-c32458306d9d\") " pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.944434 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1f4a507-b974-41a1-bdd8-c32458306d9d-serving-cert\") pod \"route-controller-manager-5df4b7c597-5xl52\" (UID: \"d1f4a507-b974-41a1-bdd8-c32458306d9d\") " pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:09 crc kubenswrapper[4815]: I0225 13:23:09.963863 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l9vk\" (UniqueName: \"kubernetes.io/projected/d1f4a507-b974-41a1-bdd8-c32458306d9d-kube-api-access-9l9vk\") pod \"route-controller-manager-5df4b7c597-5xl52\" (UID: \"d1f4a507-b974-41a1-bdd8-c32458306d9d\") " pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:10 crc kubenswrapper[4815]: I0225 13:23:10.087297 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:10 crc kubenswrapper[4815]: I0225 13:23:10.419958 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-dv7nv_65dfccf6-aa91-41fa-95d6-fd4778065408/kube-multus-additional-cni-plugins/0.log" Feb 25 13:23:10 crc kubenswrapper[4815]: I0225 13:23:10.420001 4815 generic.go:334] "Generic (PLEG): container finished" podID="65dfccf6-aa91-41fa-95d6-fd4778065408" containerID="bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9" exitCode=137 Feb 25 13:23:10 crc kubenswrapper[4815]: I0225 13:23:10.420037 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" event={"ID":"65dfccf6-aa91-41fa-95d6-fd4778065408","Type":"ContainerDied","Data":"bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9"} Feb 25 13:23:10 crc kubenswrapper[4815]: I0225 13:23:10.421562 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" event={"ID":"5577840d-a3d2-4606-8e4f-4f74401a9b2f","Type":"ContainerDied","Data":"d46265344c6eda0effb02957a4a04ebbc92361d38f83913d4f3bb1d8b754fcae"} Feb 25 13:23:10 crc kubenswrapper[4815]: I0225 13:23:10.421595 4815 scope.go:117] "RemoveContainer" containerID="b324e157d608d68382517616a269450458833a82bfa6bbbeea58aa381866aecf" Feb 25 13:23:10 crc kubenswrapper[4815]: I0225 13:23:10.421610 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f" Feb 25 13:23:10 crc kubenswrapper[4815]: I0225 13:23:10.425331 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" event={"ID":"482c5aac-8d62-4d76-8b31-ac99c1cfc4e8","Type":"ContainerDied","Data":"73ed31567e34138350e582d640ccb5a5d6ec79dffd8ecc62a2f1dcdc51e214bc"} Feb 25 13:23:10 crc kubenswrapper[4815]: I0225 13:23:10.425474 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d" Feb 25 13:23:10 crc kubenswrapper[4815]: E0225 13:23:10.426118 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29533762-lkzv9" podUID="6403af09-14ef-49fc-858f-06c1bbadb88b" Feb 25 13:23:10 crc kubenswrapper[4815]: I0225 13:23:10.451279 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f"] Feb 25 13:23:10 crc kubenswrapper[4815]: I0225 13:23:10.455096 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-c7ccfc5c7-kfv8f"] Feb 25 13:23:10 crc kubenswrapper[4815]: I0225 13:23:10.462252 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d"] Feb 25 13:23:10 crc kubenswrapper[4815]: I0225 13:23:10.464722 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bfc5fbfb9-dgt7d"] Feb 25 13:23:10 crc kubenswrapper[4815]: I0225 13:23:10.972334 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="482c5aac-8d62-4d76-8b31-ac99c1cfc4e8" path="/var/lib/kubelet/pods/482c5aac-8d62-4d76-8b31-ac99c1cfc4e8/volumes" Feb 25 13:23:10 crc kubenswrapper[4815]: I0225 13:23:10.973076 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5577840d-a3d2-4606-8e4f-4f74401a9b2f" path="/var/lib/kubelet/pods/5577840d-a3d2-4606-8e4f-4f74401a9b2f/volumes" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.583580 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-897bf449f-qd2f4"] Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.585707 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.588862 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.588929 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.589009 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.589106 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.589381 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.591369 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-897bf449f-qd2f4"] Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.610551 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.612729 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.689420 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-config\") pod \"controller-manager-897bf449f-qd2f4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.689664 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2190e107-7e6e-463e-b933-f497a368c7a4-serving-cert\") pod \"controller-manager-897bf449f-qd2f4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.689699 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-client-ca\") pod \"controller-manager-897bf449f-qd2f4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.689732 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng7nq\" (UniqueName: \"kubernetes.io/projected/2190e107-7e6e-463e-b933-f497a368c7a4-kube-api-access-ng7nq\") pod \"controller-manager-897bf449f-qd2f4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.689774 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-proxy-ca-bundles\") pod \"controller-manager-897bf449f-qd2f4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.790560 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-config\") pod \"controller-manager-897bf449f-qd2f4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.790609 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2190e107-7e6e-463e-b933-f497a368c7a4-serving-cert\") pod \"controller-manager-897bf449f-qd2f4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.790645 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-client-ca\") pod \"controller-manager-897bf449f-qd2f4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.790678 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng7nq\" (UniqueName: \"kubernetes.io/projected/2190e107-7e6e-463e-b933-f497a368c7a4-kube-api-access-ng7nq\") pod \"controller-manager-897bf449f-qd2f4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.790718 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-proxy-ca-bundles\") pod \"controller-manager-897bf449f-qd2f4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.791851 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-proxy-ca-bundles\") pod \"controller-manager-897bf449f-qd2f4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.793063 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-client-ca\") pod \"controller-manager-897bf449f-qd2f4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.793292 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-config\") pod \"controller-manager-897bf449f-qd2f4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.815163 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2190e107-7e6e-463e-b933-f497a368c7a4-serving-cert\") pod \"controller-manager-897bf449f-qd2f4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.819497 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng7nq\" (UniqueName: \"kubernetes.io/projected/2190e107-7e6e-463e-b933-f497a368c7a4-kube-api-access-ng7nq\") pod \"controller-manager-897bf449f-qd2f4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:12 crc kubenswrapper[4815]: I0225 13:23:12.911650 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:13 crc kubenswrapper[4815]: E0225 13:23:13.114725 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 25 13:23:13 crc kubenswrapper[4815]: E0225 13:23:13.114879 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5xkdr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-hbtkx_openshift-marketplace(72ffb200-965d-41f3-9900-efba5e409b8d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 25 13:23:13 crc kubenswrapper[4815]: E0225 13:23:13.116286 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-hbtkx" podUID="72ffb200-965d-41f3-9900-efba5e409b8d" Feb 25 13:23:13 crc kubenswrapper[4815]: I0225 13:23:13.678400 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-njtsc" Feb 25 13:23:13 crc kubenswrapper[4815]: E0225 13:23:13.733638 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9 is running failed: container process not found" containerID="bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 13:23:13 crc kubenswrapper[4815]: E0225 13:23:13.734678 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9 is running failed: container process not found" containerID="bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 13:23:13 crc kubenswrapper[4815]: E0225 13:23:13.735295 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9 is running failed: container process not found" containerID="bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 13:23:13 crc kubenswrapper[4815]: E0225 13:23:13.735328 4815 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9 is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" podUID="65dfccf6-aa91-41fa-95d6-fd4778065408" containerName="kube-multus-additional-cni-plugins" Feb 25 13:23:14 crc kubenswrapper[4815]: E0225 13:23:14.562074 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-hbtkx" podUID="72ffb200-965d-41f3-9900-efba5e409b8d" Feb 25 13:23:15 crc kubenswrapper[4815]: I0225 13:23:15.209865 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 25 13:23:15 crc kubenswrapper[4815]: I0225 13:23:15.211111 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 13:23:15 crc kubenswrapper[4815]: I0225 13:23:15.216653 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 25 13:23:15 crc kubenswrapper[4815]: I0225 13:23:15.217120 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 25 13:23:15 crc kubenswrapper[4815]: I0225 13:23:15.223431 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 25 13:23:15 crc kubenswrapper[4815]: I0225 13:23:15.324962 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e6412639-5ccf-44b8-9fe3-7587a3e03208-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e6412639-5ccf-44b8-9fe3-7587a3e03208\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 13:23:15 crc kubenswrapper[4815]: I0225 13:23:15.325017 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e6412639-5ccf-44b8-9fe3-7587a3e03208-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e6412639-5ccf-44b8-9fe3-7587a3e03208\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 13:23:15 crc kubenswrapper[4815]: I0225 13:23:15.426739 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e6412639-5ccf-44b8-9fe3-7587a3e03208-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e6412639-5ccf-44b8-9fe3-7587a3e03208\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 13:23:15 crc kubenswrapper[4815]: I0225 13:23:15.426790 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e6412639-5ccf-44b8-9fe3-7587a3e03208-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e6412639-5ccf-44b8-9fe3-7587a3e03208\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 13:23:15 crc kubenswrapper[4815]: I0225 13:23:15.426875 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e6412639-5ccf-44b8-9fe3-7587a3e03208-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e6412639-5ccf-44b8-9fe3-7587a3e03208\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 13:23:15 crc kubenswrapper[4815]: I0225 13:23:15.444644 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e6412639-5ccf-44b8-9fe3-7587a3e03208-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e6412639-5ccf-44b8-9fe3-7587a3e03208\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 13:23:15 crc kubenswrapper[4815]: I0225 13:23:15.537852 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 13:23:16 crc kubenswrapper[4815]: E0225 13:23:16.401338 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 25 13:23:16 crc kubenswrapper[4815]: E0225 13:23:16.402304 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rlqlg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-vzljd_openshift-marketplace(c0706f0a-7f64-47d9-b3be-afe1d760603f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 25 13:23:16 crc kubenswrapper[4815]: E0225 13:23:16.403572 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-vzljd" podUID="c0706f0a-7f64-47d9-b3be-afe1d760603f" Feb 25 13:23:16 crc kubenswrapper[4815]: E0225 13:23:16.598206 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 25 13:23:16 crc kubenswrapper[4815]: E0225 13:23:16.598338 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rvrgb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-qffvd_openshift-marketplace(2eed2795-5cbe-4dab-9e12-c9592f94befb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 25 13:23:16 crc kubenswrapper[4815]: E0225 13:23:16.599499 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-qffvd" podUID="2eed2795-5cbe-4dab-9e12-c9592f94befb" Feb 25 13:23:16 crc kubenswrapper[4815]: E0225 13:23:16.617496 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 25 13:23:16 crc kubenswrapper[4815]: E0225 13:23:16.617653 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zspjb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zm6qq_openshift-marketplace(9c7db856-98cb-47a4-bad2-35ec76191138): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 25 13:23:16 crc kubenswrapper[4815]: E0225 13:23:16.618853 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zm6qq" podUID="9c7db856-98cb-47a4-bad2-35ec76191138" Feb 25 13:23:17 crc kubenswrapper[4815]: E0225 13:23:17.850365 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-vzljd" podUID="c0706f0a-7f64-47d9-b3be-afe1d760603f" Feb 25 13:23:17 crc kubenswrapper[4815]: E0225 13:23:17.850429 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zm6qq" podUID="9c7db856-98cb-47a4-bad2-35ec76191138" Feb 25 13:23:17 crc kubenswrapper[4815]: E0225 13:23:17.852859 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-qffvd" podUID="2eed2795-5cbe-4dab-9e12-c9592f94befb" Feb 25 13:23:17 crc kubenswrapper[4815]: I0225 13:23:17.899403 4815 scope.go:117] "RemoveContainer" containerID="8f12b2e55ea7cc8e5842c756e5112c363ec72a99595bbe8616262b8642c38a4e" Feb 25 13:23:17 crc kubenswrapper[4815]: E0225 13:23:17.915277 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 25 13:23:17 crc kubenswrapper[4815]: E0225 13:23:17.915401 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fw8ct,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-x6zz5_openshift-marketplace(6d00f8b4-1bf6-4a47-9beb-2226f26d695b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 25 13:23:17 crc kubenswrapper[4815]: E0225 13:23:17.917905 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-x6zz5" podUID="6d00f8b4-1bf6-4a47-9beb-2226f26d695b" Feb 25 13:23:17 crc kubenswrapper[4815]: E0225 13:23:17.942343 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 25 13:23:17 crc kubenswrapper[4815]: E0225 13:23:17.942496 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cz6fn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-9pjrm_openshift-marketplace(bdf40b79-9414-4062-8dfd-3f0f71fcd1e4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 25 13:23:17 crc kubenswrapper[4815]: E0225 13:23:17.945566 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-9pjrm" podUID="bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" Feb 25 13:23:17 crc kubenswrapper[4815]: E0225 13:23:17.961683 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 25 13:23:17 crc kubenswrapper[4815]: E0225 13:23:17.961856 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z8vps,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-96mhl_openshift-marketplace(931648f3-732f-4bb1-81d7-6c71f2bf2971): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 25 13:23:17 crc kubenswrapper[4815]: E0225 13:23:17.963186 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-96mhl" podUID="931648f3-732f-4bb1-81d7-6c71f2bf2971" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.006957 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-dv7nv_65dfccf6-aa91-41fa-95d6-fd4778065408/kube-multus-additional-cni-plugins/0.log" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.007572 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.113287 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52"] Feb 25 13:23:18 crc kubenswrapper[4815]: W0225 13:23:18.122061 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1f4a507_b974_41a1_bdd8_c32458306d9d.slice/crio-6479d4d61951e29015f2f3794c0789d156dfd806ce9b4cad5d483369c3be067b WatchSource:0}: Error finding container 6479d4d61951e29015f2f3794c0789d156dfd806ce9b4cad5d483369c3be067b: Status 404 returned error can't find the container with id 6479d4d61951e29015f2f3794c0789d156dfd806ce9b4cad5d483369c3be067b Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.163092 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggzb8\" (UniqueName: \"kubernetes.io/projected/65dfccf6-aa91-41fa-95d6-fd4778065408-kube-api-access-ggzb8\") pod \"65dfccf6-aa91-41fa-95d6-fd4778065408\" (UID: \"65dfccf6-aa91-41fa-95d6-fd4778065408\") " Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.163127 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/65dfccf6-aa91-41fa-95d6-fd4778065408-cni-sysctl-allowlist\") pod \"65dfccf6-aa91-41fa-95d6-fd4778065408\" (UID: \"65dfccf6-aa91-41fa-95d6-fd4778065408\") " Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.163240 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/65dfccf6-aa91-41fa-95d6-fd4778065408-tuning-conf-dir\") pod \"65dfccf6-aa91-41fa-95d6-fd4778065408\" (UID: \"65dfccf6-aa91-41fa-95d6-fd4778065408\") " Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.163274 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/65dfccf6-aa91-41fa-95d6-fd4778065408-ready\") pod \"65dfccf6-aa91-41fa-95d6-fd4778065408\" (UID: \"65dfccf6-aa91-41fa-95d6-fd4778065408\") " Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.163361 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65dfccf6-aa91-41fa-95d6-fd4778065408-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "65dfccf6-aa91-41fa-95d6-fd4778065408" (UID: "65dfccf6-aa91-41fa-95d6-fd4778065408"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.163673 4815 reconciler_common.go:293] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/65dfccf6-aa91-41fa-95d6-fd4778065408-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.163738 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65dfccf6-aa91-41fa-95d6-fd4778065408-ready" (OuterVolumeSpecName: "ready") pod "65dfccf6-aa91-41fa-95d6-fd4778065408" (UID: "65dfccf6-aa91-41fa-95d6-fd4778065408"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.164002 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65dfccf6-aa91-41fa-95d6-fd4778065408-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "65dfccf6-aa91-41fa-95d6-fd4778065408" (UID: "65dfccf6-aa91-41fa-95d6-fd4778065408"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.168474 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65dfccf6-aa91-41fa-95d6-fd4778065408-kube-api-access-ggzb8" (OuterVolumeSpecName: "kube-api-access-ggzb8") pod "65dfccf6-aa91-41fa-95d6-fd4778065408" (UID: "65dfccf6-aa91-41fa-95d6-fd4778065408"). InnerVolumeSpecName "kube-api-access-ggzb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.264970 4815 reconciler_common.go:293] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/65dfccf6-aa91-41fa-95d6-fd4778065408-ready\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.265431 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggzb8\" (UniqueName: \"kubernetes.io/projected/65dfccf6-aa91-41fa-95d6-fd4778065408-kube-api-access-ggzb8\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.265447 4815 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/65dfccf6-aa91-41fa-95d6-fd4778065408-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.403764 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-897bf449f-qd2f4"] Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.408437 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.485905 4815 generic.go:334] "Generic (PLEG): container finished" podID="23366473-92d4-4958-bd1c-b6a2fb188502" containerID="07ad57f5c14711cf3ed8701d5254978f01802e812c60afc9e7248a1d45696259" exitCode=0 Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.485968 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkbjm" event={"ID":"23366473-92d4-4958-bd1c-b6a2fb188502","Type":"ContainerDied","Data":"07ad57f5c14711cf3ed8701d5254978f01802e812c60afc9e7248a1d45696259"} Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.499899 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e6412639-5ccf-44b8-9fe3-7587a3e03208","Type":"ContainerStarted","Data":"b0b382f60d59cdc71d7ccbcf18b504be9433b0a9e9f0b3c27ffcf1a7ed10e424"} Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.503236 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-dv7nv_65dfccf6-aa91-41fa-95d6-fd4778065408/kube-multus-additional-cni-plugins/0.log" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.503386 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.503776 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-dv7nv" event={"ID":"65dfccf6-aa91-41fa-95d6-fd4778065408","Type":"ContainerDied","Data":"7fda2a9d3d2ceaabbf2867e57f172d7adbcc17c7e03f299a6fdc2377d63b8c90"} Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.503820 4815 scope.go:117] "RemoveContainer" containerID="bbc7df2cbc41173da52908ecaa5c9600a5a0d76ceb59a1b126be8fb4af69adc9" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.509577 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" event={"ID":"d1f4a507-b974-41a1-bdd8-c32458306d9d","Type":"ContainerStarted","Data":"5cf7948abcc8ec4cbbe73581a3273db8ba65172b2971454a304e98f63e5fc3a7"} Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.509622 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" event={"ID":"d1f4a507-b974-41a1-bdd8-c32458306d9d","Type":"ContainerStarted","Data":"6479d4d61951e29015f2f3794c0789d156dfd806ce9b4cad5d483369c3be067b"} Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.509775 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.512076 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" event={"ID":"2190e107-7e6e-463e-b933-f497a368c7a4","Type":"ContainerStarted","Data":"654baa83041652d05b2443ac2f46fa75ea8c3ad1a9c58dcd1eb41796e3f4e594"} Feb 25 13:23:18 crc kubenswrapper[4815]: E0225 13:23:18.516754 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-96mhl" podUID="931648f3-732f-4bb1-81d7-6c71f2bf2971" Feb 25 13:23:18 crc kubenswrapper[4815]: E0225 13:23:18.517051 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-x6zz5" podUID="6d00f8b4-1bf6-4a47-9beb-2226f26d695b" Feb 25 13:23:18 crc kubenswrapper[4815]: E0225 13:23:18.517122 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-9pjrm" podUID="bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.538259 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" podStartSLOduration=19.538240376 podStartE2EDuration="19.538240376s" podCreationTimestamp="2026-02-25 13:22:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:23:18.537376246 +0000 UTC m=+156.338474300" watchObservedRunningTime="2026-02-25 13:23:18.538240376 +0000 UTC m=+156.339338430" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.602851 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-dv7nv"] Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.606565 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-dv7nv"] Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.711464 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:18 crc kubenswrapper[4815]: I0225 13:23:18.943340 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65dfccf6-aa91-41fa-95d6-fd4778065408" path="/var/lib/kubelet/pods/65dfccf6-aa91-41fa-95d6-fd4778065408/volumes" Feb 25 13:23:19 crc kubenswrapper[4815]: I0225 13:23:19.361261 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-897bf449f-qd2f4"] Feb 25 13:23:19 crc kubenswrapper[4815]: I0225 13:23:19.463852 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52"] Feb 25 13:23:19 crc kubenswrapper[4815]: I0225 13:23:19.521780 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" event={"ID":"2190e107-7e6e-463e-b933-f497a368c7a4","Type":"ContainerStarted","Data":"c7c557bef2c75ef7d8be0d114052c671f5ee3a70fcac25ef50653cd47d70d554"} Feb 25 13:23:19 crc kubenswrapper[4815]: I0225 13:23:19.522936 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:19 crc kubenswrapper[4815]: I0225 13:23:19.527605 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:19 crc kubenswrapper[4815]: I0225 13:23:19.528326 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkbjm" event={"ID":"23366473-92d4-4958-bd1c-b6a2fb188502","Type":"ContainerStarted","Data":"f30ab752ba659598d365138a94a62cf79868f2b112b936d0ed5c975ab63056cd"} Feb 25 13:23:19 crc kubenswrapper[4815]: I0225 13:23:19.529740 4815 generic.go:334] "Generic (PLEG): container finished" podID="e6412639-5ccf-44b8-9fe3-7587a3e03208" containerID="05a718197514c2d5f10c8416533fc8eb9616eb29feab64445f46e53203742ca4" exitCode=0 Feb 25 13:23:19 crc kubenswrapper[4815]: I0225 13:23:19.529783 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e6412639-5ccf-44b8-9fe3-7587a3e03208","Type":"ContainerDied","Data":"05a718197514c2d5f10c8416533fc8eb9616eb29feab64445f46e53203742ca4"} Feb 25 13:23:19 crc kubenswrapper[4815]: I0225 13:23:19.543118 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" podStartSLOduration=20.543099945 podStartE2EDuration="20.543099945s" podCreationTimestamp="2026-02-25 13:22:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:23:19.539864384 +0000 UTC m=+157.340962448" watchObservedRunningTime="2026-02-25 13:23:19.543099945 +0000 UTC m=+157.344198009" Feb 25 13:23:19 crc kubenswrapper[4815]: I0225 13:23:19.558925 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bkbjm" podStartSLOduration=3.603396119 podStartE2EDuration="40.558897227s" podCreationTimestamp="2026-02-25 13:22:39 +0000 UTC" firstStartedPulling="2026-02-25 13:22:41.933949692 +0000 UTC m=+119.735047746" lastFinishedPulling="2026-02-25 13:23:18.8894508 +0000 UTC m=+156.690548854" observedRunningTime="2026-02-25 13:23:19.556296538 +0000 UTC m=+157.357394612" watchObservedRunningTime="2026-02-25 13:23:19.558897227 +0000 UTC m=+157.359995281" Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.086677 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.325051 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bkbjm" Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.325315 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bkbjm" Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.534149 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" podUID="d1f4a507-b974-41a1-bdd8-c32458306d9d" containerName="route-controller-manager" containerID="cri-o://5cf7948abcc8ec4cbbe73581a3273db8ba65172b2971454a304e98f63e5fc3a7" gracePeriod=30 Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.534677 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" podUID="2190e107-7e6e-463e-b933-f497a368c7a4" containerName="controller-manager" containerID="cri-o://c7c557bef2c75ef7d8be0d114052c671f5ee3a70fcac25ef50653cd47d70d554" gracePeriod=30 Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.822481 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.897677 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e6412639-5ccf-44b8-9fe3-7587a3e03208-kubelet-dir\") pod \"e6412639-5ccf-44b8-9fe3-7587a3e03208\" (UID: \"e6412639-5ccf-44b8-9fe3-7587a3e03208\") " Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.897734 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e6412639-5ccf-44b8-9fe3-7587a3e03208-kube-api-access\") pod \"e6412639-5ccf-44b8-9fe3-7587a3e03208\" (UID: \"e6412639-5ccf-44b8-9fe3-7587a3e03208\") " Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.897824 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e6412639-5ccf-44b8-9fe3-7587a3e03208-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e6412639-5ccf-44b8-9fe3-7587a3e03208" (UID: "e6412639-5ccf-44b8-9fe3-7587a3e03208"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.897944 4815 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e6412639-5ccf-44b8-9fe3-7587a3e03208-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.904225 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6412639-5ccf-44b8-9fe3-7587a3e03208-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e6412639-5ccf-44b8-9fe3-7587a3e03208" (UID: "e6412639-5ccf-44b8-9fe3-7587a3e03208"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.904814 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.911052 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.999035 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1f4a507-b974-41a1-bdd8-c32458306d9d-config\") pod \"d1f4a507-b974-41a1-bdd8-c32458306d9d\" (UID: \"d1f4a507-b974-41a1-bdd8-c32458306d9d\") " Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.999088 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ng7nq\" (UniqueName: \"kubernetes.io/projected/2190e107-7e6e-463e-b933-f497a368c7a4-kube-api-access-ng7nq\") pod \"2190e107-7e6e-463e-b933-f497a368c7a4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.999116 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-config\") pod \"2190e107-7e6e-463e-b933-f497a368c7a4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.999131 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-proxy-ca-bundles\") pod \"2190e107-7e6e-463e-b933-f497a368c7a4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.999147 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-client-ca\") pod \"2190e107-7e6e-463e-b933-f497a368c7a4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.999164 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d1f4a507-b974-41a1-bdd8-c32458306d9d-client-ca\") pod \"d1f4a507-b974-41a1-bdd8-c32458306d9d\" (UID: \"d1f4a507-b974-41a1-bdd8-c32458306d9d\") " Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.999188 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9l9vk\" (UniqueName: \"kubernetes.io/projected/d1f4a507-b974-41a1-bdd8-c32458306d9d-kube-api-access-9l9vk\") pod \"d1f4a507-b974-41a1-bdd8-c32458306d9d\" (UID: \"d1f4a507-b974-41a1-bdd8-c32458306d9d\") " Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.999232 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1f4a507-b974-41a1-bdd8-c32458306d9d-serving-cert\") pod \"d1f4a507-b974-41a1-bdd8-c32458306d9d\" (UID: \"d1f4a507-b974-41a1-bdd8-c32458306d9d\") " Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.999284 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2190e107-7e6e-463e-b933-f497a368c7a4-serving-cert\") pod \"2190e107-7e6e-463e-b933-f497a368c7a4\" (UID: \"2190e107-7e6e-463e-b933-f497a368c7a4\") " Feb 25 13:23:20 crc kubenswrapper[4815]: I0225 13:23:20.999487 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e6412639-5ccf-44b8-9fe3-7587a3e03208-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.000521 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1f4a507-b974-41a1-bdd8-c32458306d9d-client-ca" (OuterVolumeSpecName: "client-ca") pod "d1f4a507-b974-41a1-bdd8-c32458306d9d" (UID: "d1f4a507-b974-41a1-bdd8-c32458306d9d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.000801 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-config" (OuterVolumeSpecName: "config") pod "2190e107-7e6e-463e-b933-f497a368c7a4" (UID: "2190e107-7e6e-463e-b933-f497a368c7a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.001035 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1f4a507-b974-41a1-bdd8-c32458306d9d-config" (OuterVolumeSpecName: "config") pod "d1f4a507-b974-41a1-bdd8-c32458306d9d" (UID: "d1f4a507-b974-41a1-bdd8-c32458306d9d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.001461 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-client-ca" (OuterVolumeSpecName: "client-ca") pod "2190e107-7e6e-463e-b933-f497a368c7a4" (UID: "2190e107-7e6e-463e-b933-f497a368c7a4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.001481 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "2190e107-7e6e-463e-b933-f497a368c7a4" (UID: "2190e107-7e6e-463e-b933-f497a368c7a4"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.002638 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2190e107-7e6e-463e-b933-f497a368c7a4-kube-api-access-ng7nq" (OuterVolumeSpecName: "kube-api-access-ng7nq") pod "2190e107-7e6e-463e-b933-f497a368c7a4" (UID: "2190e107-7e6e-463e-b933-f497a368c7a4"). InnerVolumeSpecName "kube-api-access-ng7nq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.003605 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1f4a507-b974-41a1-bdd8-c32458306d9d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d1f4a507-b974-41a1-bdd8-c32458306d9d" (UID: "d1f4a507-b974-41a1-bdd8-c32458306d9d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.003940 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2190e107-7e6e-463e-b933-f497a368c7a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2190e107-7e6e-463e-b933-f497a368c7a4" (UID: "2190e107-7e6e-463e-b933-f497a368c7a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.004116 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1f4a507-b974-41a1-bdd8-c32458306d9d-kube-api-access-9l9vk" (OuterVolumeSpecName: "kube-api-access-9l9vk") pod "d1f4a507-b974-41a1-bdd8-c32458306d9d" (UID: "d1f4a507-b974-41a1-bdd8-c32458306d9d"). InnerVolumeSpecName "kube-api-access-9l9vk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.101104 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1f4a507-b974-41a1-bdd8-c32458306d9d-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.101144 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ng7nq\" (UniqueName: \"kubernetes.io/projected/2190e107-7e6e-463e-b933-f497a368c7a4-kube-api-access-ng7nq\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.101155 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.101164 4815 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.101173 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2190e107-7e6e-463e-b933-f497a368c7a4-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.101181 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d1f4a507-b974-41a1-bdd8-c32458306d9d-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.101191 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9l9vk\" (UniqueName: \"kubernetes.io/projected/d1f4a507-b974-41a1-bdd8-c32458306d9d-kube-api-access-9l9vk\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.101201 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1f4a507-b974-41a1-bdd8-c32458306d9d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.101209 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2190e107-7e6e-463e-b933-f497a368c7a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.456227 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-bkbjm" podUID="23366473-92d4-4958-bd1c-b6a2fb188502" containerName="registry-server" probeResult="failure" output=< Feb 25 13:23:21 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Feb 25 13:23:21 crc kubenswrapper[4815]: > Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.541358 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e6412639-5ccf-44b8-9fe3-7587a3e03208","Type":"ContainerDied","Data":"b0b382f60d59cdc71d7ccbcf18b504be9433b0a9e9f0b3c27ffcf1a7ed10e424"} Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.541381 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.541433 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0b382f60d59cdc71d7ccbcf18b504be9433b0a9e9f0b3c27ffcf1a7ed10e424" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.544322 4815 generic.go:334] "Generic (PLEG): container finished" podID="d1f4a507-b974-41a1-bdd8-c32458306d9d" containerID="5cf7948abcc8ec4cbbe73581a3273db8ba65172b2971454a304e98f63e5fc3a7" exitCode=0 Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.544398 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" event={"ID":"d1f4a507-b974-41a1-bdd8-c32458306d9d","Type":"ContainerDied","Data":"5cf7948abcc8ec4cbbe73581a3273db8ba65172b2971454a304e98f63e5fc3a7"} Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.544422 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" event={"ID":"d1f4a507-b974-41a1-bdd8-c32458306d9d","Type":"ContainerDied","Data":"6479d4d61951e29015f2f3794c0789d156dfd806ce9b4cad5d483369c3be067b"} Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.544425 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.544442 4815 scope.go:117] "RemoveContainer" containerID="5cf7948abcc8ec4cbbe73581a3273db8ba65172b2971454a304e98f63e5fc3a7" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.549330 4815 generic.go:334] "Generic (PLEG): container finished" podID="2190e107-7e6e-463e-b933-f497a368c7a4" containerID="c7c557bef2c75ef7d8be0d114052c671f5ee3a70fcac25ef50653cd47d70d554" exitCode=0 Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.549391 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.549444 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" event={"ID":"2190e107-7e6e-463e-b933-f497a368c7a4","Type":"ContainerDied","Data":"c7c557bef2c75ef7d8be0d114052c671f5ee3a70fcac25ef50653cd47d70d554"} Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.549470 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-897bf449f-qd2f4" event={"ID":"2190e107-7e6e-463e-b933-f497a368c7a4","Type":"ContainerDied","Data":"654baa83041652d05b2443ac2f46fa75ea8c3ad1a9c58dcd1eb41796e3f4e594"} Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.568318 4815 scope.go:117] "RemoveContainer" containerID="5cf7948abcc8ec4cbbe73581a3273db8ba65172b2971454a304e98f63e5fc3a7" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.568635 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52"] Feb 25 13:23:21 crc kubenswrapper[4815]: E0225 13:23:21.568771 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cf7948abcc8ec4cbbe73581a3273db8ba65172b2971454a304e98f63e5fc3a7\": container with ID starting with 5cf7948abcc8ec4cbbe73581a3273db8ba65172b2971454a304e98f63e5fc3a7 not found: ID does not exist" containerID="5cf7948abcc8ec4cbbe73581a3273db8ba65172b2971454a304e98f63e5fc3a7" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.568804 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cf7948abcc8ec4cbbe73581a3273db8ba65172b2971454a304e98f63e5fc3a7"} err="failed to get container status \"5cf7948abcc8ec4cbbe73581a3273db8ba65172b2971454a304e98f63e5fc3a7\": rpc error: code = NotFound desc = could not find container \"5cf7948abcc8ec4cbbe73581a3273db8ba65172b2971454a304e98f63e5fc3a7\": container with ID starting with 5cf7948abcc8ec4cbbe73581a3273db8ba65172b2971454a304e98f63e5fc3a7 not found: ID does not exist" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.568829 4815 scope.go:117] "RemoveContainer" containerID="c7c557bef2c75ef7d8be0d114052c671f5ee3a70fcac25ef50653cd47d70d554" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.572109 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5df4b7c597-5xl52"] Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.585703 4815 scope.go:117] "RemoveContainer" containerID="c7c557bef2c75ef7d8be0d114052c671f5ee3a70fcac25ef50653cd47d70d554" Feb 25 13:23:21 crc kubenswrapper[4815]: E0225 13:23:21.586337 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7c557bef2c75ef7d8be0d114052c671f5ee3a70fcac25ef50653cd47d70d554\": container with ID starting with c7c557bef2c75ef7d8be0d114052c671f5ee3a70fcac25ef50653cd47d70d554 not found: ID does not exist" containerID="c7c557bef2c75ef7d8be0d114052c671f5ee3a70fcac25ef50653cd47d70d554" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.586366 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7c557bef2c75ef7d8be0d114052c671f5ee3a70fcac25ef50653cd47d70d554"} err="failed to get container status \"c7c557bef2c75ef7d8be0d114052c671f5ee3a70fcac25ef50653cd47d70d554\": rpc error: code = NotFound desc = could not find container \"c7c557bef2c75ef7d8be0d114052c671f5ee3a70fcac25ef50653cd47d70d554\": container with ID starting with c7c557bef2c75ef7d8be0d114052c671f5ee3a70fcac25ef50653cd47d70d554 not found: ID does not exist" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.586656 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-897bf449f-qd2f4"] Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.590021 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-897bf449f-qd2f4"] Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.603064 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 25 13:23:21 crc kubenswrapper[4815]: E0225 13:23:21.603254 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65dfccf6-aa91-41fa-95d6-fd4778065408" containerName="kube-multus-additional-cni-plugins" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.603266 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="65dfccf6-aa91-41fa-95d6-fd4778065408" containerName="kube-multus-additional-cni-plugins" Feb 25 13:23:21 crc kubenswrapper[4815]: E0225 13:23:21.603281 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f4a507-b974-41a1-bdd8-c32458306d9d" containerName="route-controller-manager" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.603287 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f4a507-b974-41a1-bdd8-c32458306d9d" containerName="route-controller-manager" Feb 25 13:23:21 crc kubenswrapper[4815]: E0225 13:23:21.603298 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6412639-5ccf-44b8-9fe3-7587a3e03208" containerName="pruner" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.603304 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6412639-5ccf-44b8-9fe3-7587a3e03208" containerName="pruner" Feb 25 13:23:21 crc kubenswrapper[4815]: E0225 13:23:21.603318 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2190e107-7e6e-463e-b933-f497a368c7a4" containerName="controller-manager" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.603324 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2190e107-7e6e-463e-b933-f497a368c7a4" containerName="controller-manager" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.603406 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6412639-5ccf-44b8-9fe3-7587a3e03208" containerName="pruner" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.603414 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="2190e107-7e6e-463e-b933-f497a368c7a4" containerName="controller-manager" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.603469 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="65dfccf6-aa91-41fa-95d6-fd4778065408" containerName="kube-multus-additional-cni-plugins" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.603478 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1f4a507-b974-41a1-bdd8-c32458306d9d" containerName="route-controller-manager" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.603834 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.607958 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.610429 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.621819 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.709414 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/276236cd-c94d-4d4b-87e6-e333a1602548-var-lock\") pod \"installer-9-crc\" (UID: \"276236cd-c94d-4d4b-87e6-e333a1602548\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.709603 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/276236cd-c94d-4d4b-87e6-e333a1602548-kube-api-access\") pod \"installer-9-crc\" (UID: \"276236cd-c94d-4d4b-87e6-e333a1602548\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.709682 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/276236cd-c94d-4d4b-87e6-e333a1602548-kubelet-dir\") pod \"installer-9-crc\" (UID: \"276236cd-c94d-4d4b-87e6-e333a1602548\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.811460 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/276236cd-c94d-4d4b-87e6-e333a1602548-kube-api-access\") pod \"installer-9-crc\" (UID: \"276236cd-c94d-4d4b-87e6-e333a1602548\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.811530 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/276236cd-c94d-4d4b-87e6-e333a1602548-kubelet-dir\") pod \"installer-9-crc\" (UID: \"276236cd-c94d-4d4b-87e6-e333a1602548\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.811605 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/276236cd-c94d-4d4b-87e6-e333a1602548-var-lock\") pod \"installer-9-crc\" (UID: \"276236cd-c94d-4d4b-87e6-e333a1602548\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.811694 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/276236cd-c94d-4d4b-87e6-e333a1602548-var-lock\") pod \"installer-9-crc\" (UID: \"276236cd-c94d-4d4b-87e6-e333a1602548\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.811695 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/276236cd-c94d-4d4b-87e6-e333a1602548-kubelet-dir\") pod \"installer-9-crc\" (UID: \"276236cd-c94d-4d4b-87e6-e333a1602548\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.828475 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/276236cd-c94d-4d4b-87e6-e333a1602548-kube-api-access\") pod \"installer-9-crc\" (UID: \"276236cd-c94d-4d4b-87e6-e333a1602548\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 13:23:21 crc kubenswrapper[4815]: I0225 13:23:21.930675 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 25 13:23:22 crc kubenswrapper[4815]: I0225 13:23:22.311571 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 25 13:23:22 crc kubenswrapper[4815]: I0225 13:23:22.557167 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"276236cd-c94d-4d4b-87e6-e333a1602548","Type":"ContainerStarted","Data":"d9b2f34bd19e97f149bc776d740d686d329196d8ee7f5d7ad452db22d7f75a77"} Feb 25 13:23:22 crc kubenswrapper[4815]: I0225 13:23:22.943730 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2190e107-7e6e-463e-b933-f497a368c7a4" path="/var/lib/kubelet/pods/2190e107-7e6e-463e-b933-f497a368c7a4/volumes" Feb 25 13:23:22 crc kubenswrapper[4815]: I0225 13:23:22.944632 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1f4a507-b974-41a1-bdd8-c32458306d9d" path="/var/lib/kubelet/pods/d1f4a507-b974-41a1-bdd8-c32458306d9d/volumes" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.566109 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"276236cd-c94d-4d4b-87e6-e333a1602548","Type":"ContainerStarted","Data":"de057a7e5b17a30c00200fb6cedfdb6137cd5db6b3d2919dda0c71f2b988c80b"} Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.596632 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg"] Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.597770 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.600154 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.600440 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.600621 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.601491 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m"] Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.601596 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.602411 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.603291 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.603270947 podStartE2EDuration="2.603270947s" podCreationTimestamp="2026-02-25 13:23:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:23:23.584807895 +0000 UTC m=+161.385905949" watchObservedRunningTime="2026-02-25 13:23:23.603270947 +0000 UTC m=+161.404369001" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.607795 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.608243 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.608285 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.608391 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.608472 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.608546 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.608977 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.609020 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.615049 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.624341 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg"] Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.629263 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m"] Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.735986 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0192d631-cd78-43f6-93f5-e73875606c6e-client-ca\") pod \"route-controller-manager-7968bf755c-qfx7m\" (UID: \"0192d631-cd78-43f6-93f5-e73875606c6e\") " pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.736021 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-proxy-ca-bundles\") pod \"controller-manager-b49d4c4bc-rwrtg\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.736044 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-client-ca\") pod \"controller-manager-b49d4c4bc-rwrtg\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.736066 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr7d2\" (UniqueName: \"kubernetes.io/projected/0192d631-cd78-43f6-93f5-e73875606c6e-kube-api-access-nr7d2\") pod \"route-controller-manager-7968bf755c-qfx7m\" (UID: \"0192d631-cd78-43f6-93f5-e73875606c6e\") " pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.736195 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-config\") pod \"controller-manager-b49d4c4bc-rwrtg\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.736292 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0192d631-cd78-43f6-93f5-e73875606c6e-serving-cert\") pod \"route-controller-manager-7968bf755c-qfx7m\" (UID: \"0192d631-cd78-43f6-93f5-e73875606c6e\") " pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.736331 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-serving-cert\") pod \"controller-manager-b49d4c4bc-rwrtg\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.736423 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0192d631-cd78-43f6-93f5-e73875606c6e-config\") pod \"route-controller-manager-7968bf755c-qfx7m\" (UID: \"0192d631-cd78-43f6-93f5-e73875606c6e\") " pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.736453 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqkgf\" (UniqueName: \"kubernetes.io/projected/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-kube-api-access-rqkgf\") pod \"controller-manager-b49d4c4bc-rwrtg\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.837171 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0192d631-cd78-43f6-93f5-e73875606c6e-config\") pod \"route-controller-manager-7968bf755c-qfx7m\" (UID: \"0192d631-cd78-43f6-93f5-e73875606c6e\") " pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.837214 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqkgf\" (UniqueName: \"kubernetes.io/projected/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-kube-api-access-rqkgf\") pod \"controller-manager-b49d4c4bc-rwrtg\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.837240 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0192d631-cd78-43f6-93f5-e73875606c6e-client-ca\") pod \"route-controller-manager-7968bf755c-qfx7m\" (UID: \"0192d631-cd78-43f6-93f5-e73875606c6e\") " pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.837258 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-proxy-ca-bundles\") pod \"controller-manager-b49d4c4bc-rwrtg\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.837282 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-client-ca\") pod \"controller-manager-b49d4c4bc-rwrtg\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.837304 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr7d2\" (UniqueName: \"kubernetes.io/projected/0192d631-cd78-43f6-93f5-e73875606c6e-kube-api-access-nr7d2\") pod \"route-controller-manager-7968bf755c-qfx7m\" (UID: \"0192d631-cd78-43f6-93f5-e73875606c6e\") " pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.837328 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-config\") pod \"controller-manager-b49d4c4bc-rwrtg\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.837369 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0192d631-cd78-43f6-93f5-e73875606c6e-serving-cert\") pod \"route-controller-manager-7968bf755c-qfx7m\" (UID: \"0192d631-cd78-43f6-93f5-e73875606c6e\") " pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.837403 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-serving-cert\") pod \"controller-manager-b49d4c4bc-rwrtg\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.838610 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0192d631-cd78-43f6-93f5-e73875606c6e-client-ca\") pod \"route-controller-manager-7968bf755c-qfx7m\" (UID: \"0192d631-cd78-43f6-93f5-e73875606c6e\") " pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.838664 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-client-ca\") pod \"controller-manager-b49d4c4bc-rwrtg\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.839143 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-proxy-ca-bundles\") pod \"controller-manager-b49d4c4bc-rwrtg\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.839186 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-config\") pod \"controller-manager-b49d4c4bc-rwrtg\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.840165 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0192d631-cd78-43f6-93f5-e73875606c6e-config\") pod \"route-controller-manager-7968bf755c-qfx7m\" (UID: \"0192d631-cd78-43f6-93f5-e73875606c6e\") " pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.842988 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-serving-cert\") pod \"controller-manager-b49d4c4bc-rwrtg\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.843131 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0192d631-cd78-43f6-93f5-e73875606c6e-serving-cert\") pod \"route-controller-manager-7968bf755c-qfx7m\" (UID: \"0192d631-cd78-43f6-93f5-e73875606c6e\") " pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.853578 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqkgf\" (UniqueName: \"kubernetes.io/projected/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-kube-api-access-rqkgf\") pod \"controller-manager-b49d4c4bc-rwrtg\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.853764 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr7d2\" (UniqueName: \"kubernetes.io/projected/0192d631-cd78-43f6-93f5-e73875606c6e-kube-api-access-nr7d2\") pod \"route-controller-manager-7968bf755c-qfx7m\" (UID: \"0192d631-cd78-43f6-93f5-e73875606c6e\") " pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.924156 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:23 crc kubenswrapper[4815]: I0225 13:23:23.944804 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:24 crc kubenswrapper[4815]: I0225 13:23:24.149901 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg"] Feb 25 13:23:24 crc kubenswrapper[4815]: I0225 13:23:24.167867 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m"] Feb 25 13:23:24 crc kubenswrapper[4815]: W0225 13:23:24.180000 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0192d631_cd78_43f6_93f5_e73875606c6e.slice/crio-ccdcf76202710cb828b43d9539c2451d466efa8a930046a9aaf64e9a1a45be8a WatchSource:0}: Error finding container ccdcf76202710cb828b43d9539c2451d466efa8a930046a9aaf64e9a1a45be8a: Status 404 returned error can't find the container with id ccdcf76202710cb828b43d9539c2451d466efa8a930046a9aaf64e9a1a45be8a Feb 25 13:23:24 crc kubenswrapper[4815]: I0225 13:23:24.572216 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" event={"ID":"0192d631-cd78-43f6-93f5-e73875606c6e","Type":"ContainerStarted","Data":"03efffc207c38425bd71ef6d16408f22eff0d45cb78ae641876a887c37d8120d"} Feb 25 13:23:24 crc kubenswrapper[4815]: I0225 13:23:24.572262 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" event={"ID":"0192d631-cd78-43f6-93f5-e73875606c6e","Type":"ContainerStarted","Data":"ccdcf76202710cb828b43d9539c2451d466efa8a930046a9aaf64e9a1a45be8a"} Feb 25 13:23:24 crc kubenswrapper[4815]: I0225 13:23:24.572435 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:24 crc kubenswrapper[4815]: I0225 13:23:24.573779 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" event={"ID":"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5","Type":"ContainerStarted","Data":"f92d975b50a315c792cf01800d0bbaf0765f69663bdf4a32665053339c50874a"} Feb 25 13:23:24 crc kubenswrapper[4815]: I0225 13:23:24.573843 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" event={"ID":"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5","Type":"ContainerStarted","Data":"37c708b3831646f65a7093c117490dc93e542e18bffa49a2052887bfef542eb1"} Feb 25 13:23:24 crc kubenswrapper[4815]: I0225 13:23:24.574150 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:24 crc kubenswrapper[4815]: I0225 13:23:24.578193 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:24 crc kubenswrapper[4815]: I0225 13:23:24.592337 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" podStartSLOduration=5.592312504 podStartE2EDuration="5.592312504s" podCreationTimestamp="2026-02-25 13:23:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:23:24.589893572 +0000 UTC m=+162.390991636" watchObservedRunningTime="2026-02-25 13:23:24.592312504 +0000 UTC m=+162.393410558" Feb 25 13:23:24 crc kubenswrapper[4815]: I0225 13:23:24.611523 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" podStartSLOduration=5.611489601 podStartE2EDuration="5.611489601s" podCreationTimestamp="2026-02-25 13:23:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:23:24.610606181 +0000 UTC m=+162.411704245" watchObservedRunningTime="2026-02-25 13:23:24.611489601 +0000 UTC m=+162.412587655" Feb 25 13:23:24 crc kubenswrapper[4815]: I0225 13:23:24.770006 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:25 crc kubenswrapper[4815]: I0225 13:23:25.492927 4815 csr.go:261] certificate signing request csr-5s8g2 is approved, waiting to be issued Feb 25 13:23:25 crc kubenswrapper[4815]: I0225 13:23:25.502313 4815 csr.go:257] certificate signing request csr-5s8g2 is issued Feb 25 13:23:25 crc kubenswrapper[4815]: I0225 13:23:25.580775 4815 generic.go:334] "Generic (PLEG): container finished" podID="6403af09-14ef-49fc-858f-06c1bbadb88b" containerID="92c3363b249ba9f79180012961834296c8953a14c98c4d67183a8e953ed394f6" exitCode=0 Feb 25 13:23:25 crc kubenswrapper[4815]: I0225 13:23:25.580868 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533762-lkzv9" event={"ID":"6403af09-14ef-49fc-858f-06c1bbadb88b","Type":"ContainerDied","Data":"92c3363b249ba9f79180012961834296c8953a14c98c4d67183a8e953ed394f6"} Feb 25 13:23:26 crc kubenswrapper[4815]: I0225 13:23:26.503558 4815 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-02 07:32:03.352588532 +0000 UTC Feb 25 13:23:26 crc kubenswrapper[4815]: I0225 13:23:26.503611 4815 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7458h8m36.848980582s for next certificate rotation Feb 25 13:23:26 crc kubenswrapper[4815]: I0225 13:23:26.828211 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533762-lkzv9" Feb 25 13:23:26 crc kubenswrapper[4815]: I0225 13:23:26.995416 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7k82t\" (UniqueName: \"kubernetes.io/projected/6403af09-14ef-49fc-858f-06c1bbadb88b-kube-api-access-7k82t\") pod \"6403af09-14ef-49fc-858f-06c1bbadb88b\" (UID: \"6403af09-14ef-49fc-858f-06c1bbadb88b\") " Feb 25 13:23:27 crc kubenswrapper[4815]: I0225 13:23:27.003913 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6403af09-14ef-49fc-858f-06c1bbadb88b-kube-api-access-7k82t" (OuterVolumeSpecName: "kube-api-access-7k82t") pod "6403af09-14ef-49fc-858f-06c1bbadb88b" (UID: "6403af09-14ef-49fc-858f-06c1bbadb88b"). InnerVolumeSpecName "kube-api-access-7k82t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:23:27 crc kubenswrapper[4815]: I0225 13:23:27.096804 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7k82t\" (UniqueName: \"kubernetes.io/projected/6403af09-14ef-49fc-858f-06c1bbadb88b-kube-api-access-7k82t\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:27 crc kubenswrapper[4815]: I0225 13:23:27.503732 4815 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-15 17:03:59.608075781 +0000 UTC Feb 25 13:23:27 crc kubenswrapper[4815]: I0225 13:23:27.504007 4815 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7035h40m32.104071065s for next certificate rotation Feb 25 13:23:27 crc kubenswrapper[4815]: I0225 13:23:27.593836 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbtkx" event={"ID":"72ffb200-965d-41f3-9900-efba5e409b8d","Type":"ContainerStarted","Data":"804ba83307795ee327042547112ff984e01258fb8dfc84734a2e1132f2ea7a17"} Feb 25 13:23:27 crc kubenswrapper[4815]: I0225 13:23:27.597050 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533762-lkzv9" event={"ID":"6403af09-14ef-49fc-858f-06c1bbadb88b","Type":"ContainerDied","Data":"25a59f5eddb13f98136269efa8e1897b0426c09398f08d2b9389d2e774e53839"} Feb 25 13:23:27 crc kubenswrapper[4815]: I0225 13:23:27.597090 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25a59f5eddb13f98136269efa8e1897b0426c09398f08d2b9389d2e774e53839" Feb 25 13:23:27 crc kubenswrapper[4815]: I0225 13:23:27.597145 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533762-lkzv9" Feb 25 13:23:28 crc kubenswrapper[4815]: I0225 13:23:28.606266 4815 generic.go:334] "Generic (PLEG): container finished" podID="72ffb200-965d-41f3-9900-efba5e409b8d" containerID="804ba83307795ee327042547112ff984e01258fb8dfc84734a2e1132f2ea7a17" exitCode=0 Feb 25 13:23:28 crc kubenswrapper[4815]: I0225 13:23:28.606304 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbtkx" event={"ID":"72ffb200-965d-41f3-9900-efba5e409b8d","Type":"ContainerDied","Data":"804ba83307795ee327042547112ff984e01258fb8dfc84734a2e1132f2ea7a17"} Feb 25 13:23:29 crc kubenswrapper[4815]: I0225 13:23:29.614392 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbtkx" event={"ID":"72ffb200-965d-41f3-9900-efba5e409b8d","Type":"ContainerStarted","Data":"9a4954e457815b32d5d802a8f265ef253d8c03154dc113ba01218287ec6b1dcd"} Feb 25 13:23:29 crc kubenswrapper[4815]: I0225 13:23:29.638121 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hbtkx" podStartSLOduration=2.818164442 podStartE2EDuration="46.638105275s" podCreationTimestamp="2026-02-25 13:22:43 +0000 UTC" firstStartedPulling="2026-02-25 13:22:45.219362806 +0000 UTC m=+123.020460860" lastFinishedPulling="2026-02-25 13:23:29.039303639 +0000 UTC m=+166.840401693" observedRunningTime="2026-02-25 13:23:29.637639639 +0000 UTC m=+167.438737693" watchObservedRunningTime="2026-02-25 13:23:29.638105275 +0000 UTC m=+167.439203329" Feb 25 13:23:30 crc kubenswrapper[4815]: I0225 13:23:30.378203 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bkbjm" Feb 25 13:23:30 crc kubenswrapper[4815]: I0225 13:23:30.427211 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bkbjm" Feb 25 13:23:31 crc kubenswrapper[4815]: I0225 13:23:31.627712 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zm6qq" event={"ID":"9c7db856-98cb-47a4-bad2-35ec76191138","Type":"ContainerStarted","Data":"4384ecf6918b5b4aafdfe77e4f0a47f7836c43b8e0751493b2d47af69beb7b9b"} Feb 25 13:23:32 crc kubenswrapper[4815]: I0225 13:23:32.636083 4815 generic.go:334] "Generic (PLEG): container finished" podID="9c7db856-98cb-47a4-bad2-35ec76191138" containerID="4384ecf6918b5b4aafdfe77e4f0a47f7836c43b8e0751493b2d47af69beb7b9b" exitCode=0 Feb 25 13:23:32 crc kubenswrapper[4815]: I0225 13:23:32.636134 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zm6qq" event={"ID":"9c7db856-98cb-47a4-bad2-35ec76191138","Type":"ContainerDied","Data":"4384ecf6918b5b4aafdfe77e4f0a47f7836c43b8e0751493b2d47af69beb7b9b"} Feb 25 13:23:32 crc kubenswrapper[4815]: I0225 13:23:32.636943 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bkbjm"] Feb 25 13:23:32 crc kubenswrapper[4815]: I0225 13:23:32.637198 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bkbjm" podUID="23366473-92d4-4958-bd1c-b6a2fb188502" containerName="registry-server" containerID="cri-o://f30ab752ba659598d365138a94a62cf79868f2b112b936d0ed5c975ab63056cd" gracePeriod=2 Feb 25 13:23:33 crc kubenswrapper[4815]: I0225 13:23:33.433543 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hbtkx" Feb 25 13:23:33 crc kubenswrapper[4815]: I0225 13:23:33.433616 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hbtkx" Feb 25 13:23:33 crc kubenswrapper[4815]: I0225 13:23:33.643278 4815 generic.go:334] "Generic (PLEG): container finished" podID="23366473-92d4-4958-bd1c-b6a2fb188502" containerID="f30ab752ba659598d365138a94a62cf79868f2b112b936d0ed5c975ab63056cd" exitCode=0 Feb 25 13:23:33 crc kubenswrapper[4815]: I0225 13:23:33.643329 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkbjm" event={"ID":"23366473-92d4-4958-bd1c-b6a2fb188502","Type":"ContainerDied","Data":"f30ab752ba659598d365138a94a62cf79868f2b112b936d0ed5c975ab63056cd"} Feb 25 13:23:34 crc kubenswrapper[4815]: I0225 13:23:34.339652 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bkbjm" Feb 25 13:23:34 crc kubenswrapper[4815]: I0225 13:23:34.489328 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23366473-92d4-4958-bd1c-b6a2fb188502-utilities\") pod \"23366473-92d4-4958-bd1c-b6a2fb188502\" (UID: \"23366473-92d4-4958-bd1c-b6a2fb188502\") " Feb 25 13:23:34 crc kubenswrapper[4815]: I0225 13:23:34.489376 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23366473-92d4-4958-bd1c-b6a2fb188502-catalog-content\") pod \"23366473-92d4-4958-bd1c-b6a2fb188502\" (UID: \"23366473-92d4-4958-bd1c-b6a2fb188502\") " Feb 25 13:23:34 crc kubenswrapper[4815]: I0225 13:23:34.489468 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwc99\" (UniqueName: \"kubernetes.io/projected/23366473-92d4-4958-bd1c-b6a2fb188502-kube-api-access-xwc99\") pod \"23366473-92d4-4958-bd1c-b6a2fb188502\" (UID: \"23366473-92d4-4958-bd1c-b6a2fb188502\") " Feb 25 13:23:34 crc kubenswrapper[4815]: I0225 13:23:34.490192 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23366473-92d4-4958-bd1c-b6a2fb188502-utilities" (OuterVolumeSpecName: "utilities") pod "23366473-92d4-4958-bd1c-b6a2fb188502" (UID: "23366473-92d4-4958-bd1c-b6a2fb188502"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:23:34 crc kubenswrapper[4815]: I0225 13:23:34.494641 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23366473-92d4-4958-bd1c-b6a2fb188502-kube-api-access-xwc99" (OuterVolumeSpecName: "kube-api-access-xwc99") pod "23366473-92d4-4958-bd1c-b6a2fb188502" (UID: "23366473-92d4-4958-bd1c-b6a2fb188502"). InnerVolumeSpecName "kube-api-access-xwc99". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:23:34 crc kubenswrapper[4815]: I0225 13:23:34.496816 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hbtkx" podUID="72ffb200-965d-41f3-9900-efba5e409b8d" containerName="registry-server" probeResult="failure" output=< Feb 25 13:23:34 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Feb 25 13:23:34 crc kubenswrapper[4815]: > Feb 25 13:23:34 crc kubenswrapper[4815]: I0225 13:23:34.547780 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23366473-92d4-4958-bd1c-b6a2fb188502-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23366473-92d4-4958-bd1c-b6a2fb188502" (UID: "23366473-92d4-4958-bd1c-b6a2fb188502"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:23:34 crc kubenswrapper[4815]: I0225 13:23:34.590411 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23366473-92d4-4958-bd1c-b6a2fb188502-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:34 crc kubenswrapper[4815]: I0225 13:23:34.590453 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23366473-92d4-4958-bd1c-b6a2fb188502-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:34 crc kubenswrapper[4815]: I0225 13:23:34.590469 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwc99\" (UniqueName: \"kubernetes.io/projected/23366473-92d4-4958-bd1c-b6a2fb188502-kube-api-access-xwc99\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:34 crc kubenswrapper[4815]: I0225 13:23:34.650979 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkbjm" event={"ID":"23366473-92d4-4958-bd1c-b6a2fb188502","Type":"ContainerDied","Data":"b4fb42e181cacc2765330219221d0bf5a53a5c1eede8dc6bfd0dc88ede6f5f8d"} Feb 25 13:23:34 crc kubenswrapper[4815]: I0225 13:23:34.651033 4815 scope.go:117] "RemoveContainer" containerID="f30ab752ba659598d365138a94a62cf79868f2b112b936d0ed5c975ab63056cd" Feb 25 13:23:34 crc kubenswrapper[4815]: I0225 13:23:34.651082 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bkbjm" Feb 25 13:23:34 crc kubenswrapper[4815]: I0225 13:23:34.677563 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bkbjm"] Feb 25 13:23:34 crc kubenswrapper[4815]: I0225 13:23:34.681580 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bkbjm"] Feb 25 13:23:34 crc kubenswrapper[4815]: I0225 13:23:34.952804 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23366473-92d4-4958-bd1c-b6a2fb188502" path="/var/lib/kubelet/pods/23366473-92d4-4958-bd1c-b6a2fb188502/volumes" Feb 25 13:23:37 crc kubenswrapper[4815]: I0225 13:23:37.673628 4815 scope.go:117] "RemoveContainer" containerID="07ad57f5c14711cf3ed8701d5254978f01802e812c60afc9e7248a1d45696259" Feb 25 13:23:37 crc kubenswrapper[4815]: I0225 13:23:37.693685 4815 scope.go:117] "RemoveContainer" containerID="37b745cb4e1580b68d2ba3d751e25df91a9fe3dee311b410061c12dd30ff8026" Feb 25 13:23:38 crc kubenswrapper[4815]: I0225 13:23:38.675096 4815 generic.go:334] "Generic (PLEG): container finished" podID="c0706f0a-7f64-47d9-b3be-afe1d760603f" containerID="901fef9ce43607b5aee6845a1d8fe41620110fa225779efaa464a24d1bcd2f55" exitCode=0 Feb 25 13:23:38 crc kubenswrapper[4815]: I0225 13:23:38.675185 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzljd" event={"ID":"c0706f0a-7f64-47d9-b3be-afe1d760603f","Type":"ContainerDied","Data":"901fef9ce43607b5aee6845a1d8fe41620110fa225779efaa464a24d1bcd2f55"} Feb 25 13:23:38 crc kubenswrapper[4815]: I0225 13:23:38.678252 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zm6qq" event={"ID":"9c7db856-98cb-47a4-bad2-35ec76191138","Type":"ContainerStarted","Data":"95e998a6ac1c1396f48b6482b190372c7309388aa8da448275af3ba3f6b9fd1a"} Feb 25 13:23:38 crc kubenswrapper[4815]: I0225 13:23:38.682249 4815 generic.go:334] "Generic (PLEG): container finished" podID="6d00f8b4-1bf6-4a47-9beb-2226f26d695b" containerID="a48ac1b49a391ebb803ae6d4669a0f80ae09a3be5f5ff2b349a1ba02e8f275fb" exitCode=0 Feb 25 13:23:38 crc kubenswrapper[4815]: I0225 13:23:38.682332 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6zz5" event={"ID":"6d00f8b4-1bf6-4a47-9beb-2226f26d695b","Type":"ContainerDied","Data":"a48ac1b49a391ebb803ae6d4669a0f80ae09a3be5f5ff2b349a1ba02e8f275fb"} Feb 25 13:23:38 crc kubenswrapper[4815]: I0225 13:23:38.684587 4815 generic.go:334] "Generic (PLEG): container finished" podID="bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" containerID="3f599ec4efa479f6413f95f683103e67ad256120d566559eded38b0a9cd092ec" exitCode=0 Feb 25 13:23:38 crc kubenswrapper[4815]: I0225 13:23:38.684643 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9pjrm" event={"ID":"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4","Type":"ContainerDied","Data":"3f599ec4efa479f6413f95f683103e67ad256120d566559eded38b0a9cd092ec"} Feb 25 13:23:38 crc kubenswrapper[4815]: I0225 13:23:38.689401 4815 generic.go:334] "Generic (PLEG): container finished" podID="2eed2795-5cbe-4dab-9e12-c9592f94befb" containerID="96dbb8be03ad2921477ffc18e552bc001a0f894e43de9611629cb09aede01dd6" exitCode=0 Feb 25 13:23:38 crc kubenswrapper[4815]: I0225 13:23:38.689471 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qffvd" event={"ID":"2eed2795-5cbe-4dab-9e12-c9592f94befb","Type":"ContainerDied","Data":"96dbb8be03ad2921477ffc18e552bc001a0f894e43de9611629cb09aede01dd6"} Feb 25 13:23:38 crc kubenswrapper[4815]: I0225 13:23:38.701664 4815 generic.go:334] "Generic (PLEG): container finished" podID="931648f3-732f-4bb1-81d7-6c71f2bf2971" containerID="ebfa03a2f33b6ad72d614f55d48a2fa3c6a7888c331e8cf4b17b9bbd7b43e3f6" exitCode=0 Feb 25 13:23:38 crc kubenswrapper[4815]: I0225 13:23:38.701747 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96mhl" event={"ID":"931648f3-732f-4bb1-81d7-6c71f2bf2971","Type":"ContainerDied","Data":"ebfa03a2f33b6ad72d614f55d48a2fa3c6a7888c331e8cf4b17b9bbd7b43e3f6"} Feb 25 13:23:38 crc kubenswrapper[4815]: I0225 13:23:38.727662 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zm6qq" podStartSLOduration=3.971088795 podStartE2EDuration="59.727640706s" podCreationTimestamp="2026-02-25 13:22:39 +0000 UTC" firstStartedPulling="2026-02-25 13:22:41.937201701 +0000 UTC m=+119.738299755" lastFinishedPulling="2026-02-25 13:23:37.693753592 +0000 UTC m=+175.494851666" observedRunningTime="2026-02-25 13:23:38.719049552 +0000 UTC m=+176.520147616" watchObservedRunningTime="2026-02-25 13:23:38.727640706 +0000 UTC m=+176.528738750" Feb 25 13:23:39 crc kubenswrapper[4815]: I0225 13:23:39.392730 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg"] Feb 25 13:23:39 crc kubenswrapper[4815]: I0225 13:23:39.393302 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" podUID="ff50fbe3-83d4-4a84-afea-4e2ec84e57d5" containerName="controller-manager" containerID="cri-o://f92d975b50a315c792cf01800d0bbaf0765f69663bdf4a32665053339c50874a" gracePeriod=30 Feb 25 13:23:39 crc kubenswrapper[4815]: I0225 13:23:39.428952 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m"] Feb 25 13:23:39 crc kubenswrapper[4815]: I0225 13:23:39.429181 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" podUID="0192d631-cd78-43f6-93f5-e73875606c6e" containerName="route-controller-manager" containerID="cri-o://03efffc207c38425bd71ef6d16408f22eff0d45cb78ae641876a887c37d8120d" gracePeriod=30 Feb 25 13:23:39 crc kubenswrapper[4815]: I0225 13:23:39.714830 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qffvd" event={"ID":"2eed2795-5cbe-4dab-9e12-c9592f94befb","Type":"ContainerStarted","Data":"87e7fc405fa7c0840afd21611903ce6ea3ce64d71f0b5ed8b8802b42019bf3ef"} Feb 25 13:23:39 crc kubenswrapper[4815]: I0225 13:23:39.716251 4815 generic.go:334] "Generic (PLEG): container finished" podID="ff50fbe3-83d4-4a84-afea-4e2ec84e57d5" containerID="f92d975b50a315c792cf01800d0bbaf0765f69663bdf4a32665053339c50874a" exitCode=0 Feb 25 13:23:39 crc kubenswrapper[4815]: I0225 13:23:39.716306 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" event={"ID":"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5","Type":"ContainerDied","Data":"f92d975b50a315c792cf01800d0bbaf0765f69663bdf4a32665053339c50874a"} Feb 25 13:23:39 crc kubenswrapper[4815]: I0225 13:23:39.722150 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96mhl" event={"ID":"931648f3-732f-4bb1-81d7-6c71f2bf2971","Type":"ContainerStarted","Data":"95b46293cf330e85c7fb349a7bc3e06ecd47998289939b9e8193777c471d0a28"} Feb 25 13:23:39 crc kubenswrapper[4815]: I0225 13:23:39.725010 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzljd" event={"ID":"c0706f0a-7f64-47d9-b3be-afe1d760603f","Type":"ContainerStarted","Data":"ba6d7bc00ccdc194c823eea650bd84480e9545de9ba1158a9d9711d07c69862f"} Feb 25 13:23:39 crc kubenswrapper[4815]: I0225 13:23:39.727611 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6zz5" event={"ID":"6d00f8b4-1bf6-4a47-9beb-2226f26d695b","Type":"ContainerStarted","Data":"4eeb3c6973f20f0a56118b5dbf28db1a25436109bf0347e578acad431a1051b9"} Feb 25 13:23:39 crc kubenswrapper[4815]: I0225 13:23:39.740694 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-96mhl" podStartSLOduration=2.804321336 podStartE2EDuration="57.740681875s" podCreationTimestamp="2026-02-25 13:22:42 +0000 UTC" firstStartedPulling="2026-02-25 13:22:44.177389396 +0000 UTC m=+121.978487450" lastFinishedPulling="2026-02-25 13:23:39.113749935 +0000 UTC m=+176.914847989" observedRunningTime="2026-02-25 13:23:39.739472133 +0000 UTC m=+177.540570187" watchObservedRunningTime="2026-02-25 13:23:39.740681875 +0000 UTC m=+177.541779929" Feb 25 13:23:39 crc kubenswrapper[4815]: I0225 13:23:39.759757 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x6zz5" podStartSLOduration=2.767684042 podStartE2EDuration="58.759735128s" podCreationTimestamp="2026-02-25 13:22:41 +0000 UTC" firstStartedPulling="2026-02-25 13:22:43.121756521 +0000 UTC m=+120.922854575" lastFinishedPulling="2026-02-25 13:23:39.113807587 +0000 UTC m=+176.914905661" observedRunningTime="2026-02-25 13:23:39.758748294 +0000 UTC m=+177.559846348" watchObservedRunningTime="2026-02-25 13:23:39.759735128 +0000 UTC m=+177.560833182" Feb 25 13:23:39 crc kubenswrapper[4815]: I0225 13:23:39.858646 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zm6qq" Feb 25 13:23:39 crc kubenswrapper[4815]: I0225 13:23:39.858689 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zm6qq" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.667123 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.695564 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg"] Feb 25 13:23:40 crc kubenswrapper[4815]: E0225 13:23:40.695778 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23366473-92d4-4958-bd1c-b6a2fb188502" containerName="registry-server" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.695795 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="23366473-92d4-4958-bd1c-b6a2fb188502" containerName="registry-server" Feb 25 13:23:40 crc kubenswrapper[4815]: E0225 13:23:40.695808 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0192d631-cd78-43f6-93f5-e73875606c6e" containerName="route-controller-manager" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.695818 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="0192d631-cd78-43f6-93f5-e73875606c6e" containerName="route-controller-manager" Feb 25 13:23:40 crc kubenswrapper[4815]: E0225 13:23:40.695827 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23366473-92d4-4958-bd1c-b6a2fb188502" containerName="extract-utilities" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.695835 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="23366473-92d4-4958-bd1c-b6a2fb188502" containerName="extract-utilities" Feb 25 13:23:40 crc kubenswrapper[4815]: E0225 13:23:40.695852 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6403af09-14ef-49fc-858f-06c1bbadb88b" containerName="oc" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.695858 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6403af09-14ef-49fc-858f-06c1bbadb88b" containerName="oc" Feb 25 13:23:40 crc kubenswrapper[4815]: E0225 13:23:40.695867 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23366473-92d4-4958-bd1c-b6a2fb188502" containerName="extract-content" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.695873 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="23366473-92d4-4958-bd1c-b6a2fb188502" containerName="extract-content" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.695976 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="0192d631-cd78-43f6-93f5-e73875606c6e" containerName="route-controller-manager" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.695991 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="23366473-92d4-4958-bd1c-b6a2fb188502" containerName="registry-server" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.696007 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="6403af09-14ef-49fc-858f-06c1bbadb88b" containerName="oc" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.696365 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.720537 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.737382 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" event={"ID":"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5","Type":"ContainerDied","Data":"37c708b3831646f65a7093c117490dc93e542e18bffa49a2052887bfef542eb1"} Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.737434 4815 scope.go:117] "RemoveContainer" containerID="f92d975b50a315c792cf01800d0bbaf0765f69663bdf4a32665053339c50874a" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.737567 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.742464 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9pjrm" event={"ID":"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4","Type":"ContainerStarted","Data":"8248af535da5869c4f84a7e1453d2b519a4ff9c0ea2de89b618c2dcb5ea1040b"} Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.743588 4815 generic.go:334] "Generic (PLEG): container finished" podID="0192d631-cd78-43f6-93f5-e73875606c6e" containerID="03efffc207c38425bd71ef6d16408f22eff0d45cb78ae641876a887c37d8120d" exitCode=0 Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.743647 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.743681 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" event={"ID":"0192d631-cd78-43f6-93f5-e73875606c6e","Type":"ContainerDied","Data":"03efffc207c38425bd71ef6d16408f22eff0d45cb78ae641876a887c37d8120d"} Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.743701 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m" event={"ID":"0192d631-cd78-43f6-93f5-e73875606c6e","Type":"ContainerDied","Data":"ccdcf76202710cb828b43d9539c2451d466efa8a930046a9aaf64e9a1a45be8a"} Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.767245 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nr7d2\" (UniqueName: \"kubernetes.io/projected/0192d631-cd78-43f6-93f5-e73875606c6e-kube-api-access-nr7d2\") pod \"0192d631-cd78-43f6-93f5-e73875606c6e\" (UID: \"0192d631-cd78-43f6-93f5-e73875606c6e\") " Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.767663 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0192d631-cd78-43f6-93f5-e73875606c6e-serving-cert\") pod \"0192d631-cd78-43f6-93f5-e73875606c6e\" (UID: \"0192d631-cd78-43f6-93f5-e73875606c6e\") " Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.767711 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0192d631-cd78-43f6-93f5-e73875606c6e-client-ca\") pod \"0192d631-cd78-43f6-93f5-e73875606c6e\" (UID: \"0192d631-cd78-43f6-93f5-e73875606c6e\") " Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.767764 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0192d631-cd78-43f6-93f5-e73875606c6e-config\") pod \"0192d631-cd78-43f6-93f5-e73875606c6e\" (UID: \"0192d631-cd78-43f6-93f5-e73875606c6e\") " Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.772243 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0192d631-cd78-43f6-93f5-e73875606c6e-config" (OuterVolumeSpecName: "config") pod "0192d631-cd78-43f6-93f5-e73875606c6e" (UID: "0192d631-cd78-43f6-93f5-e73875606c6e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.775215 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0192d631-cd78-43f6-93f5-e73875606c6e-kube-api-access-nr7d2" (OuterVolumeSpecName: "kube-api-access-nr7d2") pod "0192d631-cd78-43f6-93f5-e73875606c6e" (UID: "0192d631-cd78-43f6-93f5-e73875606c6e"). InnerVolumeSpecName "kube-api-access-nr7d2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.775276 4815 scope.go:117] "RemoveContainer" containerID="03efffc207c38425bd71ef6d16408f22eff0d45cb78ae641876a887c37d8120d" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.775988 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0192d631-cd78-43f6-93f5-e73875606c6e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0192d631-cd78-43f6-93f5-e73875606c6e" (UID: "0192d631-cd78-43f6-93f5-e73875606c6e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.778101 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0192d631-cd78-43f6-93f5-e73875606c6e-client-ca" (OuterVolumeSpecName: "client-ca") pod "0192d631-cd78-43f6-93f5-e73875606c6e" (UID: "0192d631-cd78-43f6-93f5-e73875606c6e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.825938 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9pjrm" podStartSLOduration=2.734753134 podStartE2EDuration="58.825922728s" podCreationTimestamp="2026-02-25 13:22:42 +0000 UTC" firstStartedPulling="2026-02-25 13:22:44.156328806 +0000 UTC m=+121.957426860" lastFinishedPulling="2026-02-25 13:23:40.2474984 +0000 UTC m=+178.048596454" observedRunningTime="2026-02-25 13:23:40.805156536 +0000 UTC m=+178.606254620" watchObservedRunningTime="2026-02-25 13:23:40.825922728 +0000 UTC m=+178.627020782" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.825980 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg"] Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.828198 4815 scope.go:117] "RemoveContainer" containerID="03efffc207c38425bd71ef6d16408f22eff0d45cb78ae641876a887c37d8120d" Feb 25 13:23:40 crc kubenswrapper[4815]: E0225 13:23:40.833691 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03efffc207c38425bd71ef6d16408f22eff0d45cb78ae641876a887c37d8120d\": container with ID starting with 03efffc207c38425bd71ef6d16408f22eff0d45cb78ae641876a887c37d8120d not found: ID does not exist" containerID="03efffc207c38425bd71ef6d16408f22eff0d45cb78ae641876a887c37d8120d" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.833743 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03efffc207c38425bd71ef6d16408f22eff0d45cb78ae641876a887c37d8120d"} err="failed to get container status \"03efffc207c38425bd71ef6d16408f22eff0d45cb78ae641876a887c37d8120d\": rpc error: code = NotFound desc = could not find container \"03efffc207c38425bd71ef6d16408f22eff0d45cb78ae641876a887c37d8120d\": container with ID starting with 03efffc207c38425bd71ef6d16408f22eff0d45cb78ae641876a887c37d8120d not found: ID does not exist" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.869911 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-client-ca\") pod \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.869967 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-proxy-ca-bundles\") pod \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.870053 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-config\") pod \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.870079 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-serving-cert\") pod \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.870112 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqkgf\" (UniqueName: \"kubernetes.io/projected/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-kube-api-access-rqkgf\") pod \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\" (UID: \"ff50fbe3-83d4-4a84-afea-4e2ec84e57d5\") " Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.870298 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40e8b948-118a-404c-8683-3a800e159f03-serving-cert\") pod \"route-controller-manager-54767cd7bb-vdcdg\" (UID: \"40e8b948-118a-404c-8683-3a800e159f03\") " pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.870342 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40e8b948-118a-404c-8683-3a800e159f03-config\") pod \"route-controller-manager-54767cd7bb-vdcdg\" (UID: \"40e8b948-118a-404c-8683-3a800e159f03\") " pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.870437 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40e8b948-118a-404c-8683-3a800e159f03-client-ca\") pod \"route-controller-manager-54767cd7bb-vdcdg\" (UID: \"40e8b948-118a-404c-8683-3a800e159f03\") " pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.870494 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4r69\" (UniqueName: \"kubernetes.io/projected/40e8b948-118a-404c-8683-3a800e159f03-kube-api-access-x4r69\") pod \"route-controller-manager-54767cd7bb-vdcdg\" (UID: \"40e8b948-118a-404c-8683-3a800e159f03\") " pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.870587 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nr7d2\" (UniqueName: \"kubernetes.io/projected/0192d631-cd78-43f6-93f5-e73875606c6e-kube-api-access-nr7d2\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.870601 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0192d631-cd78-43f6-93f5-e73875606c6e-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.870612 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0192d631-cd78-43f6-93f5-e73875606c6e-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.870622 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0192d631-cd78-43f6-93f5-e73875606c6e-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.871439 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ff50fbe3-83d4-4a84-afea-4e2ec84e57d5" (UID: "ff50fbe3-83d4-4a84-afea-4e2ec84e57d5"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.869922 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qffvd" podStartSLOduration=3.625440776 podStartE2EDuration="1m0.869887984s" podCreationTimestamp="2026-02-25 13:22:40 +0000 UTC" firstStartedPulling="2026-02-25 13:22:41.972375678 +0000 UTC m=+119.773473732" lastFinishedPulling="2026-02-25 13:23:39.216822876 +0000 UTC m=+177.017920940" observedRunningTime="2026-02-25 13:23:40.846929078 +0000 UTC m=+178.648027142" watchObservedRunningTime="2026-02-25 13:23:40.869887984 +0000 UTC m=+178.670986038" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.871795 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-client-ca" (OuterVolumeSpecName: "client-ca") pod "ff50fbe3-83d4-4a84-afea-4e2ec84e57d5" (UID: "ff50fbe3-83d4-4a84-afea-4e2ec84e57d5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.872422 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-config" (OuterVolumeSpecName: "config") pod "ff50fbe3-83d4-4a84-afea-4e2ec84e57d5" (UID: "ff50fbe3-83d4-4a84-afea-4e2ec84e57d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.873258 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vzljd" podStartSLOduration=4.615576886 podStartE2EDuration="1m1.873249199s" podCreationTimestamp="2026-02-25 13:22:39 +0000 UTC" firstStartedPulling="2026-02-25 13:22:41.960054885 +0000 UTC m=+119.761152939" lastFinishedPulling="2026-02-25 13:23:39.217727188 +0000 UTC m=+177.018825252" observedRunningTime="2026-02-25 13:23:40.868710764 +0000 UTC m=+178.669808838" watchObservedRunningTime="2026-02-25 13:23:40.873249199 +0000 UTC m=+178.674347253" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.878146 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-kube-api-access-rqkgf" (OuterVolumeSpecName: "kube-api-access-rqkgf") pod "ff50fbe3-83d4-4a84-afea-4e2ec84e57d5" (UID: "ff50fbe3-83d4-4a84-afea-4e2ec84e57d5"). InnerVolumeSpecName "kube-api-access-rqkgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.886665 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ff50fbe3-83d4-4a84-afea-4e2ec84e57d5" (UID: "ff50fbe3-83d4-4a84-afea-4e2ec84e57d5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.901973 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-zm6qq" podUID="9c7db856-98cb-47a4-bad2-35ec76191138" containerName="registry-server" probeResult="failure" output=< Feb 25 13:23:40 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Feb 25 13:23:40 crc kubenswrapper[4815]: > Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.971843 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40e8b948-118a-404c-8683-3a800e159f03-serving-cert\") pod \"route-controller-manager-54767cd7bb-vdcdg\" (UID: \"40e8b948-118a-404c-8683-3a800e159f03\") " pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.971904 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40e8b948-118a-404c-8683-3a800e159f03-config\") pod \"route-controller-manager-54767cd7bb-vdcdg\" (UID: \"40e8b948-118a-404c-8683-3a800e159f03\") " pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.971961 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40e8b948-118a-404c-8683-3a800e159f03-client-ca\") pod \"route-controller-manager-54767cd7bb-vdcdg\" (UID: \"40e8b948-118a-404c-8683-3a800e159f03\") " pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.971992 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4r69\" (UniqueName: \"kubernetes.io/projected/40e8b948-118a-404c-8683-3a800e159f03-kube-api-access-x4r69\") pod \"route-controller-manager-54767cd7bb-vdcdg\" (UID: \"40e8b948-118a-404c-8683-3a800e159f03\") " pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.972030 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.972041 4815 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.972053 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.972061 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.972070 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqkgf\" (UniqueName: \"kubernetes.io/projected/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5-kube-api-access-rqkgf\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.973116 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40e8b948-118a-404c-8683-3a800e159f03-client-ca\") pod \"route-controller-manager-54767cd7bb-vdcdg\" (UID: \"40e8b948-118a-404c-8683-3a800e159f03\") " pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.973322 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40e8b948-118a-404c-8683-3a800e159f03-config\") pod \"route-controller-manager-54767cd7bb-vdcdg\" (UID: \"40e8b948-118a-404c-8683-3a800e159f03\") " pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.980247 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40e8b948-118a-404c-8683-3a800e159f03-serving-cert\") pod \"route-controller-manager-54767cd7bb-vdcdg\" (UID: \"40e8b948-118a-404c-8683-3a800e159f03\") " pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:23:40 crc kubenswrapper[4815]: I0225 13:23:40.994799 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4r69\" (UniqueName: \"kubernetes.io/projected/40e8b948-118a-404c-8683-3a800e159f03-kube-api-access-x4r69\") pod \"route-controller-manager-54767cd7bb-vdcdg\" (UID: \"40e8b948-118a-404c-8683-3a800e159f03\") " pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:23:41 crc kubenswrapper[4815]: I0225 13:23:41.031228 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:23:41 crc kubenswrapper[4815]: I0225 13:23:41.059033 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg"] Feb 25 13:23:41 crc kubenswrapper[4815]: I0225 13:23:41.062918 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-b49d4c4bc-rwrtg"] Feb 25 13:23:41 crc kubenswrapper[4815]: I0225 13:23:41.070839 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m"] Feb 25 13:23:41 crc kubenswrapper[4815]: I0225 13:23:41.076447 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7968bf755c-qfx7m"] Feb 25 13:23:41 crc kubenswrapper[4815]: I0225 13:23:41.458469 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg"] Feb 25 13:23:41 crc kubenswrapper[4815]: I0225 13:23:41.752220 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" event={"ID":"40e8b948-118a-404c-8683-3a800e159f03","Type":"ContainerStarted","Data":"fc9e0681c3d6bd9c78d893b8852443115a507df82241b05734da777a03a3a18d"} Feb 25 13:23:41 crc kubenswrapper[4815]: I0225 13:23:41.752470 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" event={"ID":"40e8b948-118a-404c-8683-3a800e159f03","Type":"ContainerStarted","Data":"1b9ce5bb39c93526ab45d127201a8be6d42d87396c057800794632bd7d251b27"} Feb 25 13:23:41 crc kubenswrapper[4815]: I0225 13:23:41.752854 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:23:41 crc kubenswrapper[4815]: I0225 13:23:41.776463 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" podStartSLOduration=2.776444096 podStartE2EDuration="2.776444096s" podCreationTimestamp="2026-02-25 13:23:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:23:41.776185687 +0000 UTC m=+179.577283751" watchObservedRunningTime="2026-02-25 13:23:41.776444096 +0000 UTC m=+179.577542150" Feb 25 13:23:42 crc kubenswrapper[4815]: I0225 13:23:42.010097 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x6zz5" Feb 25 13:23:42 crc kubenswrapper[4815]: I0225 13:23:42.010451 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x6zz5" Feb 25 13:23:42 crc kubenswrapper[4815]: I0225 13:23:42.052436 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:23:42 crc kubenswrapper[4815]: I0225 13:23:42.062956 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x6zz5" Feb 25 13:23:42 crc kubenswrapper[4815]: I0225 13:23:42.428809 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-96mhl" Feb 25 13:23:42 crc kubenswrapper[4815]: I0225 13:23:42.428890 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-96mhl" Feb 25 13:23:42 crc kubenswrapper[4815]: I0225 13:23:42.496364 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-96mhl" Feb 25 13:23:42 crc kubenswrapper[4815]: I0225 13:23:42.949723 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0192d631-cd78-43f6-93f5-e73875606c6e" path="/var/lib/kubelet/pods/0192d631-cd78-43f6-93f5-e73875606c6e/volumes" Feb 25 13:23:42 crc kubenswrapper[4815]: I0225 13:23:42.951087 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff50fbe3-83d4-4a84-afea-4e2ec84e57d5" path="/var/lib/kubelet/pods/ff50fbe3-83d4-4a84-afea-4e2ec84e57d5/volumes" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.022309 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9pjrm" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.022387 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9pjrm" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.504014 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hbtkx" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.562555 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hbtkx" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.599901 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-74f875b898-j2txv"] Feb 25 13:23:43 crc kubenswrapper[4815]: E0225 13:23:43.600217 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff50fbe3-83d4-4a84-afea-4e2ec84e57d5" containerName="controller-manager" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.600245 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff50fbe3-83d4-4a84-afea-4e2ec84e57d5" containerName="controller-manager" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.600402 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff50fbe3-83d4-4a84-afea-4e2ec84e57d5" containerName="controller-manager" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.601039 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.603317 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.603355 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.603485 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.605342 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.606037 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.606129 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.609536 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74f875b898-j2txv"] Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.617141 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.705592 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-config\") pod \"controller-manager-74f875b898-j2txv\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.705749 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-client-ca\") pod \"controller-manager-74f875b898-j2txv\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.705812 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-proxy-ca-bundles\") pod \"controller-manager-74f875b898-j2txv\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.705948 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnvcl\" (UniqueName: \"kubernetes.io/projected/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-kube-api-access-lnvcl\") pod \"controller-manager-74f875b898-j2txv\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.705979 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-serving-cert\") pod \"controller-manager-74f875b898-j2txv\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.806844 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnvcl\" (UniqueName: \"kubernetes.io/projected/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-kube-api-access-lnvcl\") pod \"controller-manager-74f875b898-j2txv\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.806920 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-serving-cert\") pod \"controller-manager-74f875b898-j2txv\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.806947 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-config\") pod \"controller-manager-74f875b898-j2txv\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.807840 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-client-ca\") pod \"controller-manager-74f875b898-j2txv\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.807868 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-proxy-ca-bundles\") pod \"controller-manager-74f875b898-j2txv\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.808550 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-config\") pod \"controller-manager-74f875b898-j2txv\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.809295 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-client-ca\") pod \"controller-manager-74f875b898-j2txv\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.809401 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-proxy-ca-bundles\") pod \"controller-manager-74f875b898-j2txv\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.814919 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-serving-cert\") pod \"controller-manager-74f875b898-j2txv\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.839097 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnvcl\" (UniqueName: \"kubernetes.io/projected/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-kube-api-access-lnvcl\") pod \"controller-manager-74f875b898-j2txv\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:43 crc kubenswrapper[4815]: I0225 13:23:43.927826 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:44 crc kubenswrapper[4815]: I0225 13:23:44.087657 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9pjrm" podUID="bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" containerName="registry-server" probeResult="failure" output=< Feb 25 13:23:44 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Feb 25 13:23:44 crc kubenswrapper[4815]: > Feb 25 13:23:44 crc kubenswrapper[4815]: I0225 13:23:44.115398 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74f875b898-j2txv"] Feb 25 13:23:44 crc kubenswrapper[4815]: I0225 13:23:44.772224 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" event={"ID":"9d9c770e-abbd-4c20-a96f-bf95d2a89d28","Type":"ContainerStarted","Data":"d9596a73a7255f1be66892c03c316906f3fa30994be69afc49392561e3465cea"} Feb 25 13:23:44 crc kubenswrapper[4815]: I0225 13:23:44.772618 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" event={"ID":"9d9c770e-abbd-4c20-a96f-bf95d2a89d28","Type":"ContainerStarted","Data":"259cada00403507ac09e8fd220104ddf3f0720762595a91a9a4ad629667b46ed"} Feb 25 13:23:45 crc kubenswrapper[4815]: I0225 13:23:45.432839 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hbtkx"] Feb 25 13:23:45 crc kubenswrapper[4815]: I0225 13:23:45.433165 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hbtkx" podUID="72ffb200-965d-41f3-9900-efba5e409b8d" containerName="registry-server" containerID="cri-o://9a4954e457815b32d5d802a8f265ef253d8c03154dc113ba01218287ec6b1dcd" gracePeriod=2 Feb 25 13:23:45 crc kubenswrapper[4815]: I0225 13:23:45.781090 4815 generic.go:334] "Generic (PLEG): container finished" podID="72ffb200-965d-41f3-9900-efba5e409b8d" containerID="9a4954e457815b32d5d802a8f265ef253d8c03154dc113ba01218287ec6b1dcd" exitCode=0 Feb 25 13:23:45 crc kubenswrapper[4815]: I0225 13:23:45.781155 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbtkx" event={"ID":"72ffb200-965d-41f3-9900-efba5e409b8d","Type":"ContainerDied","Data":"9a4954e457815b32d5d802a8f265ef253d8c03154dc113ba01218287ec6b1dcd"} Feb 25 13:23:45 crc kubenswrapper[4815]: I0225 13:23:45.781818 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:45 crc kubenswrapper[4815]: I0225 13:23:45.785556 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:23:45 crc kubenswrapper[4815]: I0225 13:23:45.804788 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" podStartSLOduration=6.8047671770000004 podStartE2EDuration="6.804767177s" podCreationTimestamp="2026-02-25 13:23:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:23:45.801246046 +0000 UTC m=+183.602344110" watchObservedRunningTime="2026-02-25 13:23:45.804767177 +0000 UTC m=+183.605865231" Feb 25 13:23:45 crc kubenswrapper[4815]: I0225 13:23:45.871100 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hbtkx" Feb 25 13:23:45 crc kubenswrapper[4815]: I0225 13:23:45.949359 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72ffb200-965d-41f3-9900-efba5e409b8d-utilities\") pod \"72ffb200-965d-41f3-9900-efba5e409b8d\" (UID: \"72ffb200-965d-41f3-9900-efba5e409b8d\") " Feb 25 13:23:45 crc kubenswrapper[4815]: I0225 13:23:45.949491 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72ffb200-965d-41f3-9900-efba5e409b8d-catalog-content\") pod \"72ffb200-965d-41f3-9900-efba5e409b8d\" (UID: \"72ffb200-965d-41f3-9900-efba5e409b8d\") " Feb 25 13:23:45 crc kubenswrapper[4815]: I0225 13:23:45.949605 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xkdr\" (UniqueName: \"kubernetes.io/projected/72ffb200-965d-41f3-9900-efba5e409b8d-kube-api-access-5xkdr\") pod \"72ffb200-965d-41f3-9900-efba5e409b8d\" (UID: \"72ffb200-965d-41f3-9900-efba5e409b8d\") " Feb 25 13:23:45 crc kubenswrapper[4815]: I0225 13:23:45.950721 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72ffb200-965d-41f3-9900-efba5e409b8d-utilities" (OuterVolumeSpecName: "utilities") pod "72ffb200-965d-41f3-9900-efba5e409b8d" (UID: "72ffb200-965d-41f3-9900-efba5e409b8d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:23:45 crc kubenswrapper[4815]: I0225 13:23:45.955339 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72ffb200-965d-41f3-9900-efba5e409b8d-kube-api-access-5xkdr" (OuterVolumeSpecName: "kube-api-access-5xkdr") pod "72ffb200-965d-41f3-9900-efba5e409b8d" (UID: "72ffb200-965d-41f3-9900-efba5e409b8d"). InnerVolumeSpecName "kube-api-access-5xkdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:23:46 crc kubenswrapper[4815]: I0225 13:23:46.051169 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xkdr\" (UniqueName: \"kubernetes.io/projected/72ffb200-965d-41f3-9900-efba5e409b8d-kube-api-access-5xkdr\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:46 crc kubenswrapper[4815]: I0225 13:23:46.051209 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72ffb200-965d-41f3-9900-efba5e409b8d-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:46 crc kubenswrapper[4815]: I0225 13:23:46.084687 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72ffb200-965d-41f3-9900-efba5e409b8d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72ffb200-965d-41f3-9900-efba5e409b8d" (UID: "72ffb200-965d-41f3-9900-efba5e409b8d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:23:46 crc kubenswrapper[4815]: I0225 13:23:46.152411 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72ffb200-965d-41f3-9900-efba5e409b8d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:46 crc kubenswrapper[4815]: I0225 13:23:46.791177 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hbtkx" event={"ID":"72ffb200-965d-41f3-9900-efba5e409b8d","Type":"ContainerDied","Data":"b581c3a933ce725d120845df267555a60feddd8f3e7317026c06e7850d765c14"} Feb 25 13:23:46 crc kubenswrapper[4815]: I0225 13:23:46.791220 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hbtkx" Feb 25 13:23:46 crc kubenswrapper[4815]: I0225 13:23:46.791237 4815 scope.go:117] "RemoveContainer" containerID="9a4954e457815b32d5d802a8f265ef253d8c03154dc113ba01218287ec6b1dcd" Feb 25 13:23:46 crc kubenswrapper[4815]: I0225 13:23:46.806069 4815 scope.go:117] "RemoveContainer" containerID="804ba83307795ee327042547112ff984e01258fb8dfc84734a2e1132f2ea7a17" Feb 25 13:23:46 crc kubenswrapper[4815]: I0225 13:23:46.824946 4815 scope.go:117] "RemoveContainer" containerID="d8aa12ff096ee354b371f67ee8a4bb21f9987ee44fa58067b111355b0c744e06" Feb 25 13:23:46 crc kubenswrapper[4815]: I0225 13:23:46.830333 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hbtkx"] Feb 25 13:23:46 crc kubenswrapper[4815]: I0225 13:23:46.842956 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hbtkx"] Feb 25 13:23:46 crc kubenswrapper[4815]: I0225 13:23:46.941724 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72ffb200-965d-41f3-9900-efba5e409b8d" path="/var/lib/kubelet/pods/72ffb200-965d-41f3-9900-efba5e409b8d/volumes" Feb 25 13:23:49 crc kubenswrapper[4815]: I0225 13:23:49.918343 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zm6qq" Feb 25 13:23:49 crc kubenswrapper[4815]: I0225 13:23:49.963821 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zm6qq" Feb 25 13:23:50 crc kubenswrapper[4815]: I0225 13:23:50.026738 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vzljd" Feb 25 13:23:50 crc kubenswrapper[4815]: I0225 13:23:50.026797 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vzljd" Feb 25 13:23:50 crc kubenswrapper[4815]: I0225 13:23:50.062600 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vzljd" Feb 25 13:23:50 crc kubenswrapper[4815]: I0225 13:23:50.419603 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qffvd" Feb 25 13:23:50 crc kubenswrapper[4815]: I0225 13:23:50.419652 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qffvd" Feb 25 13:23:50 crc kubenswrapper[4815]: I0225 13:23:50.468108 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qffvd" Feb 25 13:23:50 crc kubenswrapper[4815]: I0225 13:23:50.876204 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qffvd" Feb 25 13:23:50 crc kubenswrapper[4815]: I0225 13:23:50.885411 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vzljd" Feb 25 13:23:52 crc kubenswrapper[4815]: I0225 13:23:52.033225 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qffvd"] Feb 25 13:23:52 crc kubenswrapper[4815]: I0225 13:23:52.072409 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x6zz5" Feb 25 13:23:52 crc kubenswrapper[4815]: I0225 13:23:52.474405 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-96mhl" Feb 25 13:23:52 crc kubenswrapper[4815]: I0225 13:23:52.835700 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qffvd" podUID="2eed2795-5cbe-4dab-9e12-c9592f94befb" containerName="registry-server" containerID="cri-o://87e7fc405fa7c0840afd21611903ce6ea3ce64d71f0b5ed8b8802b42019bf3ef" gracePeriod=2 Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.067228 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9pjrm" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.131831 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9pjrm" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.391004 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qffvd" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.466296 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2eed2795-5cbe-4dab-9e12-c9592f94befb-catalog-content\") pod \"2eed2795-5cbe-4dab-9e12-c9592f94befb\" (UID: \"2eed2795-5cbe-4dab-9e12-c9592f94befb\") " Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.466373 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvrgb\" (UniqueName: \"kubernetes.io/projected/2eed2795-5cbe-4dab-9e12-c9592f94befb-kube-api-access-rvrgb\") pod \"2eed2795-5cbe-4dab-9e12-c9592f94befb\" (UID: \"2eed2795-5cbe-4dab-9e12-c9592f94befb\") " Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.466417 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2eed2795-5cbe-4dab-9e12-c9592f94befb-utilities\") pod \"2eed2795-5cbe-4dab-9e12-c9592f94befb\" (UID: \"2eed2795-5cbe-4dab-9e12-c9592f94befb\") " Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.468254 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2eed2795-5cbe-4dab-9e12-c9592f94befb-utilities" (OuterVolumeSpecName: "utilities") pod "2eed2795-5cbe-4dab-9e12-c9592f94befb" (UID: "2eed2795-5cbe-4dab-9e12-c9592f94befb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.472810 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eed2795-5cbe-4dab-9e12-c9592f94befb-kube-api-access-rvrgb" (OuterVolumeSpecName: "kube-api-access-rvrgb") pod "2eed2795-5cbe-4dab-9e12-c9592f94befb" (UID: "2eed2795-5cbe-4dab-9e12-c9592f94befb"). InnerVolumeSpecName "kube-api-access-rvrgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.541077 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2eed2795-5cbe-4dab-9e12-c9592f94befb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2eed2795-5cbe-4dab-9e12-c9592f94befb" (UID: "2eed2795-5cbe-4dab-9e12-c9592f94befb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.567445 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvrgb\" (UniqueName: \"kubernetes.io/projected/2eed2795-5cbe-4dab-9e12-c9592f94befb-kube-api-access-rvrgb\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.567481 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2eed2795-5cbe-4dab-9e12-c9592f94befb-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.567495 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2eed2795-5cbe-4dab-9e12-c9592f94befb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.845617 4815 generic.go:334] "Generic (PLEG): container finished" podID="2eed2795-5cbe-4dab-9e12-c9592f94befb" containerID="87e7fc405fa7c0840afd21611903ce6ea3ce64d71f0b5ed8b8802b42019bf3ef" exitCode=0 Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.845744 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qffvd" event={"ID":"2eed2795-5cbe-4dab-9e12-c9592f94befb","Type":"ContainerDied","Data":"87e7fc405fa7c0840afd21611903ce6ea3ce64d71f0b5ed8b8802b42019bf3ef"} Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.845858 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qffvd" event={"ID":"2eed2795-5cbe-4dab-9e12-c9592f94befb","Type":"ContainerDied","Data":"67ba65baa8bfafdbb506dfc3eea5e0feb3ae3a47a7e21eed55d24585282cc6aa"} Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.845902 4815 scope.go:117] "RemoveContainer" containerID="87e7fc405fa7c0840afd21611903ce6ea3ce64d71f0b5ed8b8802b42019bf3ef" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.845790 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qffvd" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.875350 4815 scope.go:117] "RemoveContainer" containerID="96dbb8be03ad2921477ffc18e552bc001a0f894e43de9611629cb09aede01dd6" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.908797 4815 scope.go:117] "RemoveContainer" containerID="5c55e82bf68f55bf6a6aa089e5a131267568f93fed4e8cd6c8782d2bf23a26ab" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.909539 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qffvd"] Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.916059 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qffvd"] Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.936638 4815 scope.go:117] "RemoveContainer" containerID="87e7fc405fa7c0840afd21611903ce6ea3ce64d71f0b5ed8b8802b42019bf3ef" Feb 25 13:23:53 crc kubenswrapper[4815]: E0225 13:23:53.937141 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87e7fc405fa7c0840afd21611903ce6ea3ce64d71f0b5ed8b8802b42019bf3ef\": container with ID starting with 87e7fc405fa7c0840afd21611903ce6ea3ce64d71f0b5ed8b8802b42019bf3ef not found: ID does not exist" containerID="87e7fc405fa7c0840afd21611903ce6ea3ce64d71f0b5ed8b8802b42019bf3ef" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.937229 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87e7fc405fa7c0840afd21611903ce6ea3ce64d71f0b5ed8b8802b42019bf3ef"} err="failed to get container status \"87e7fc405fa7c0840afd21611903ce6ea3ce64d71f0b5ed8b8802b42019bf3ef\": rpc error: code = NotFound desc = could not find container \"87e7fc405fa7c0840afd21611903ce6ea3ce64d71f0b5ed8b8802b42019bf3ef\": container with ID starting with 87e7fc405fa7c0840afd21611903ce6ea3ce64d71f0b5ed8b8802b42019bf3ef not found: ID does not exist" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.937341 4815 scope.go:117] "RemoveContainer" containerID="96dbb8be03ad2921477ffc18e552bc001a0f894e43de9611629cb09aede01dd6" Feb 25 13:23:53 crc kubenswrapper[4815]: E0225 13:23:53.937735 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96dbb8be03ad2921477ffc18e552bc001a0f894e43de9611629cb09aede01dd6\": container with ID starting with 96dbb8be03ad2921477ffc18e552bc001a0f894e43de9611629cb09aede01dd6 not found: ID does not exist" containerID="96dbb8be03ad2921477ffc18e552bc001a0f894e43de9611629cb09aede01dd6" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.937851 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96dbb8be03ad2921477ffc18e552bc001a0f894e43de9611629cb09aede01dd6"} err="failed to get container status \"96dbb8be03ad2921477ffc18e552bc001a0f894e43de9611629cb09aede01dd6\": rpc error: code = NotFound desc = could not find container \"96dbb8be03ad2921477ffc18e552bc001a0f894e43de9611629cb09aede01dd6\": container with ID starting with 96dbb8be03ad2921477ffc18e552bc001a0f894e43de9611629cb09aede01dd6 not found: ID does not exist" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.937956 4815 scope.go:117] "RemoveContainer" containerID="5c55e82bf68f55bf6a6aa089e5a131267568f93fed4e8cd6c8782d2bf23a26ab" Feb 25 13:23:53 crc kubenswrapper[4815]: E0225 13:23:53.938250 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c55e82bf68f55bf6a6aa089e5a131267568f93fed4e8cd6c8782d2bf23a26ab\": container with ID starting with 5c55e82bf68f55bf6a6aa089e5a131267568f93fed4e8cd6c8782d2bf23a26ab not found: ID does not exist" containerID="5c55e82bf68f55bf6a6aa089e5a131267568f93fed4e8cd6c8782d2bf23a26ab" Feb 25 13:23:53 crc kubenswrapper[4815]: I0225 13:23:53.938338 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c55e82bf68f55bf6a6aa089e5a131267568f93fed4e8cd6c8782d2bf23a26ab"} err="failed to get container status \"5c55e82bf68f55bf6a6aa089e5a131267568f93fed4e8cd6c8782d2bf23a26ab\": rpc error: code = NotFound desc = could not find container \"5c55e82bf68f55bf6a6aa089e5a131267568f93fed4e8cd6c8782d2bf23a26ab\": container with ID starting with 5c55e82bf68f55bf6a6aa089e5a131267568f93fed4e8cd6c8782d2bf23a26ab not found: ID does not exist" Feb 25 13:23:54 crc kubenswrapper[4815]: I0225 13:23:54.633324 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-96mhl"] Feb 25 13:23:54 crc kubenswrapper[4815]: I0225 13:23:54.634887 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-96mhl" podUID="931648f3-732f-4bb1-81d7-6c71f2bf2971" containerName="registry-server" containerID="cri-o://95b46293cf330e85c7fb349a7bc3e06ecd47998289939b9e8193777c471d0a28" gracePeriod=2 Feb 25 13:23:54 crc kubenswrapper[4815]: I0225 13:23:54.863933 4815 generic.go:334] "Generic (PLEG): container finished" podID="931648f3-732f-4bb1-81d7-6c71f2bf2971" containerID="95b46293cf330e85c7fb349a7bc3e06ecd47998289939b9e8193777c471d0a28" exitCode=0 Feb 25 13:23:54 crc kubenswrapper[4815]: I0225 13:23:54.864023 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96mhl" event={"ID":"931648f3-732f-4bb1-81d7-6c71f2bf2971","Type":"ContainerDied","Data":"95b46293cf330e85c7fb349a7bc3e06ecd47998289939b9e8193777c471d0a28"} Feb 25 13:23:54 crc kubenswrapper[4815]: I0225 13:23:54.949850 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2eed2795-5cbe-4dab-9e12-c9592f94befb" path="/var/lib/kubelet/pods/2eed2795-5cbe-4dab-9e12-c9592f94befb/volumes" Feb 25 13:23:54 crc kubenswrapper[4815]: I0225 13:23:54.993252 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-md7bn"] Feb 25 13:23:55 crc kubenswrapper[4815]: I0225 13:23:55.166438 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96mhl" Feb 25 13:23:55 crc kubenswrapper[4815]: I0225 13:23:55.293079 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/931648f3-732f-4bb1-81d7-6c71f2bf2971-utilities\") pod \"931648f3-732f-4bb1-81d7-6c71f2bf2971\" (UID: \"931648f3-732f-4bb1-81d7-6c71f2bf2971\") " Feb 25 13:23:55 crc kubenswrapper[4815]: I0225 13:23:55.293566 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8vps\" (UniqueName: \"kubernetes.io/projected/931648f3-732f-4bb1-81d7-6c71f2bf2971-kube-api-access-z8vps\") pod \"931648f3-732f-4bb1-81d7-6c71f2bf2971\" (UID: \"931648f3-732f-4bb1-81d7-6c71f2bf2971\") " Feb 25 13:23:55 crc kubenswrapper[4815]: I0225 13:23:55.293631 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/931648f3-732f-4bb1-81d7-6c71f2bf2971-catalog-content\") pod \"931648f3-732f-4bb1-81d7-6c71f2bf2971\" (UID: \"931648f3-732f-4bb1-81d7-6c71f2bf2971\") " Feb 25 13:23:55 crc kubenswrapper[4815]: I0225 13:23:55.293758 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/931648f3-732f-4bb1-81d7-6c71f2bf2971-utilities" (OuterVolumeSpecName: "utilities") pod "931648f3-732f-4bb1-81d7-6c71f2bf2971" (UID: "931648f3-732f-4bb1-81d7-6c71f2bf2971"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:23:55 crc kubenswrapper[4815]: I0225 13:23:55.300986 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/931648f3-732f-4bb1-81d7-6c71f2bf2971-kube-api-access-z8vps" (OuterVolumeSpecName: "kube-api-access-z8vps") pod "931648f3-732f-4bb1-81d7-6c71f2bf2971" (UID: "931648f3-732f-4bb1-81d7-6c71f2bf2971"). InnerVolumeSpecName "kube-api-access-z8vps". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:23:55 crc kubenswrapper[4815]: I0225 13:23:55.309316 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/931648f3-732f-4bb1-81d7-6c71f2bf2971-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:55 crc kubenswrapper[4815]: I0225 13:23:55.310317 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8vps\" (UniqueName: \"kubernetes.io/projected/931648f3-732f-4bb1-81d7-6c71f2bf2971-kube-api-access-z8vps\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:55 crc kubenswrapper[4815]: I0225 13:23:55.319733 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/931648f3-732f-4bb1-81d7-6c71f2bf2971-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "931648f3-732f-4bb1-81d7-6c71f2bf2971" (UID: "931648f3-732f-4bb1-81d7-6c71f2bf2971"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:23:55 crc kubenswrapper[4815]: I0225 13:23:55.412244 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/931648f3-732f-4bb1-81d7-6c71f2bf2971-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:23:55 crc kubenswrapper[4815]: I0225 13:23:55.875423 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96mhl" event={"ID":"931648f3-732f-4bb1-81d7-6c71f2bf2971","Type":"ContainerDied","Data":"1b773db05120718fa34285eba8e410e83b89ef80f85f20c8099eee116b344734"} Feb 25 13:23:55 crc kubenswrapper[4815]: I0225 13:23:55.875595 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96mhl" Feb 25 13:23:55 crc kubenswrapper[4815]: I0225 13:23:55.876618 4815 scope.go:117] "RemoveContainer" containerID="95b46293cf330e85c7fb349a7bc3e06ecd47998289939b9e8193777c471d0a28" Feb 25 13:23:55 crc kubenswrapper[4815]: I0225 13:23:55.908365 4815 scope.go:117] "RemoveContainer" containerID="ebfa03a2f33b6ad72d614f55d48a2fa3c6a7888c331e8cf4b17b9bbd7b43e3f6" Feb 25 13:23:55 crc kubenswrapper[4815]: I0225 13:23:55.921198 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-96mhl"] Feb 25 13:23:55 crc kubenswrapper[4815]: I0225 13:23:55.926309 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-96mhl"] Feb 25 13:23:55 crc kubenswrapper[4815]: I0225 13:23:55.954454 4815 scope.go:117] "RemoveContainer" containerID="cecc9c8ae4ef2e44d1df26f3cb5804514de160ff0579bfb126efdf7c2252f1a9" Feb 25 13:23:56 crc kubenswrapper[4815]: I0225 13:23:56.945709 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="931648f3-732f-4bb1-81d7-6c71f2bf2971" path="/var/lib/kubelet/pods/931648f3-732f-4bb1-81d7-6c71f2bf2971/volumes" Feb 25 13:23:59 crc kubenswrapper[4815]: I0225 13:23:59.400995 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74f875b898-j2txv"] Feb 25 13:23:59 crc kubenswrapper[4815]: I0225 13:23:59.401620 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" podUID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" containerName="controller-manager" containerID="cri-o://d9596a73a7255f1be66892c03c316906f3fa30994be69afc49392561e3465cea" gracePeriod=30 Feb 25 13:23:59 crc kubenswrapper[4815]: I0225 13:23:59.490466 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg"] Feb 25 13:23:59 crc kubenswrapper[4815]: I0225 13:23:59.490761 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" podUID="40e8b948-118a-404c-8683-3a800e159f03" containerName="route-controller-manager" containerID="cri-o://fc9e0681c3d6bd9c78d893b8852443115a507df82241b05734da777a03a3a18d" gracePeriod=30 Feb 25 13:23:59 crc kubenswrapper[4815]: W0225 13:23:59.508031 4815 container.go:586] Failed to update stats for container "/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d9c770e_abbd_4c20_a96f_bf95d2a89d28.slice/crio-259cada00403507ac09e8fd220104ddf3f0720762595a91a9a4ad629667b46ed": error while statting cgroup v2: [unable to parse /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d9c770e_abbd_4c20_a96f_bf95d2a89d28.slice/crio-259cada00403507ac09e8fd220104ddf3f0720762595a91a9a4ad629667b46ed/memory.stat: read /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d9c770e_abbd_4c20_a96f_bf95d2a89d28.slice/crio-259cada00403507ac09e8fd220104ddf3f0720762595a91a9a4ad629667b46ed/memory.stat: no such device], continuing to push stats Feb 25 13:23:59 crc kubenswrapper[4815]: I0225 13:23:59.899894 4815 generic.go:334] "Generic (PLEG): container finished" podID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" containerID="d9596a73a7255f1be66892c03c316906f3fa30994be69afc49392561e3465cea" exitCode=0 Feb 25 13:23:59 crc kubenswrapper[4815]: I0225 13:23:59.899979 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" event={"ID":"9d9c770e-abbd-4c20-a96f-bf95d2a89d28","Type":"ContainerDied","Data":"d9596a73a7255f1be66892c03c316906f3fa30994be69afc49392561e3465cea"} Feb 25 13:23:59 crc kubenswrapper[4815]: I0225 13:23:59.901848 4815 generic.go:334] "Generic (PLEG): container finished" podID="40e8b948-118a-404c-8683-3a800e159f03" containerID="fc9e0681c3d6bd9c78d893b8852443115a507df82241b05734da777a03a3a18d" exitCode=0 Feb 25 13:23:59 crc kubenswrapper[4815]: I0225 13:23:59.901882 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" event={"ID":"40e8b948-118a-404c-8683-3a800e159f03","Type":"ContainerDied","Data":"fc9e0681c3d6bd9c78d893b8852443115a507df82241b05734da777a03a3a18d"} Feb 25 13:23:59 crc kubenswrapper[4815]: I0225 13:23:59.901902 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" event={"ID":"40e8b948-118a-404c-8683-3a800e159f03","Type":"ContainerDied","Data":"1b9ce5bb39c93526ab45d127201a8be6d42d87396c057800794632bd7d251b27"} Feb 25 13:23:59 crc kubenswrapper[4815]: I0225 13:23:59.901912 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b9ce5bb39c93526ab45d127201a8be6d42d87396c057800794632bd7d251b27" Feb 25 13:23:59 crc kubenswrapper[4815]: I0225 13:23:59.929910 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:23:59 crc kubenswrapper[4815]: I0225 13:23:59.937012 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.079210 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40e8b948-118a-404c-8683-3a800e159f03-config\") pod \"40e8b948-118a-404c-8683-3a800e159f03\" (UID: \"40e8b948-118a-404c-8683-3a800e159f03\") " Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.079486 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-serving-cert\") pod \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.079617 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-config\") pod \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.079701 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnvcl\" (UniqueName: \"kubernetes.io/projected/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-kube-api-access-lnvcl\") pod \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.079777 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4r69\" (UniqueName: \"kubernetes.io/projected/40e8b948-118a-404c-8683-3a800e159f03-kube-api-access-x4r69\") pod \"40e8b948-118a-404c-8683-3a800e159f03\" (UID: \"40e8b948-118a-404c-8683-3a800e159f03\") " Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.079886 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40e8b948-118a-404c-8683-3a800e159f03-client-ca\") pod \"40e8b948-118a-404c-8683-3a800e159f03\" (UID: \"40e8b948-118a-404c-8683-3a800e159f03\") " Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.080247 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40e8b948-118a-404c-8683-3a800e159f03-config" (OuterVolumeSpecName: "config") pod "40e8b948-118a-404c-8683-3a800e159f03" (UID: "40e8b948-118a-404c-8683-3a800e159f03"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.080329 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-config" (OuterVolumeSpecName: "config") pod "9d9c770e-abbd-4c20-a96f-bf95d2a89d28" (UID: "9d9c770e-abbd-4c20-a96f-bf95d2a89d28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.080407 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40e8b948-118a-404c-8683-3a800e159f03-client-ca" (OuterVolumeSpecName: "client-ca") pod "40e8b948-118a-404c-8683-3a800e159f03" (UID: "40e8b948-118a-404c-8683-3a800e159f03"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.080925 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-client-ca\") pod \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.081259 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-proxy-ca-bundles\") pod \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\" (UID: \"9d9c770e-abbd-4c20-a96f-bf95d2a89d28\") " Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.081759 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40e8b948-118a-404c-8683-3a800e159f03-serving-cert\") pod \"40e8b948-118a-404c-8683-3a800e159f03\" (UID: \"40e8b948-118a-404c-8683-3a800e159f03\") " Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.081194 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-client-ca" (OuterVolumeSpecName: "client-ca") pod "9d9c770e-abbd-4c20-a96f-bf95d2a89d28" (UID: "9d9c770e-abbd-4c20-a96f-bf95d2a89d28"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.081711 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9d9c770e-abbd-4c20-a96f-bf95d2a89d28" (UID: "9d9c770e-abbd-4c20-a96f-bf95d2a89d28"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.082286 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.082362 4815 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.082445 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40e8b948-118a-404c-8683-3a800e159f03-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.082519 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.082578 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40e8b948-118a-404c-8683-3a800e159f03-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.084615 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e8b948-118a-404c-8683-3a800e159f03-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "40e8b948-118a-404c-8683-3a800e159f03" (UID: "40e8b948-118a-404c-8683-3a800e159f03"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.084630 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d9c770e-abbd-4c20-a96f-bf95d2a89d28" (UID: "9d9c770e-abbd-4c20-a96f-bf95d2a89d28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.084718 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40e8b948-118a-404c-8683-3a800e159f03-kube-api-access-x4r69" (OuterVolumeSpecName: "kube-api-access-x4r69") pod "40e8b948-118a-404c-8683-3a800e159f03" (UID: "40e8b948-118a-404c-8683-3a800e159f03"). InnerVolumeSpecName "kube-api-access-x4r69". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.084897 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-kube-api-access-lnvcl" (OuterVolumeSpecName: "kube-api-access-lnvcl") pod "9d9c770e-abbd-4c20-a96f-bf95d2a89d28" (UID: "9d9c770e-abbd-4c20-a96f-bf95d2a89d28"). InnerVolumeSpecName "kube-api-access-lnvcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.128962 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533764-9mvkb"] Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.129210 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" containerName="controller-manager" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.129222 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" containerName="controller-manager" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.129238 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eed2795-5cbe-4dab-9e12-c9592f94befb" containerName="extract-utilities" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.129245 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eed2795-5cbe-4dab-9e12-c9592f94befb" containerName="extract-utilities" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.129253 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ffb200-965d-41f3-9900-efba5e409b8d" containerName="extract-utilities" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.129260 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ffb200-965d-41f3-9900-efba5e409b8d" containerName="extract-utilities" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.129270 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="931648f3-732f-4bb1-81d7-6c71f2bf2971" containerName="extract-content" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.129275 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="931648f3-732f-4bb1-81d7-6c71f2bf2971" containerName="extract-content" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.129282 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ffb200-965d-41f3-9900-efba5e409b8d" containerName="registry-server" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.129288 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ffb200-965d-41f3-9900-efba5e409b8d" containerName="registry-server" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.129296 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ffb200-965d-41f3-9900-efba5e409b8d" containerName="extract-content" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.129302 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ffb200-965d-41f3-9900-efba5e409b8d" containerName="extract-content" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.129308 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="931648f3-732f-4bb1-81d7-6c71f2bf2971" containerName="registry-server" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.129313 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="931648f3-732f-4bb1-81d7-6c71f2bf2971" containerName="registry-server" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.129319 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40e8b948-118a-404c-8683-3a800e159f03" containerName="route-controller-manager" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.129326 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="40e8b948-118a-404c-8683-3a800e159f03" containerName="route-controller-manager" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.129334 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eed2795-5cbe-4dab-9e12-c9592f94befb" containerName="extract-content" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.129340 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eed2795-5cbe-4dab-9e12-c9592f94befb" containerName="extract-content" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.129350 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eed2795-5cbe-4dab-9e12-c9592f94befb" containerName="registry-server" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.129355 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eed2795-5cbe-4dab-9e12-c9592f94befb" containerName="registry-server" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.129365 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="931648f3-732f-4bb1-81d7-6c71f2bf2971" containerName="extract-utilities" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.129371 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="931648f3-732f-4bb1-81d7-6c71f2bf2971" containerName="extract-utilities" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.129463 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eed2795-5cbe-4dab-9e12-c9592f94befb" containerName="registry-server" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.129473 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" containerName="controller-manager" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.129479 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="40e8b948-118a-404c-8683-3a800e159f03" containerName="route-controller-manager" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.129487 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="931648f3-732f-4bb1-81d7-6c71f2bf2971" containerName="registry-server" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.129539 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="72ffb200-965d-41f3-9900-efba5e409b8d" containerName="registry-server" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.129923 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533764-9mvkb" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.134530 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533764-9mvkb"] Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.134888 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.134907 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.135069 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.153115 4815 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.153452 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://fa8220d4ac2c87c277e7e2a77c5d619bfb5af22a635106397a1c4104d420a70c" gracePeriod=15 Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.153458 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://38d1ca633d56103d9a2620de025cf43520585a65f1f225944a7a2f02714c8600" gracePeriod=15 Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.153519 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://f25f11a9ab520446a86fd17fe4c1ffc9bf0622c1bbf72019dbb1535caba77279" gracePeriod=15 Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.153445 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://75f8f6e705409d40c4115c32c4e36f96c90c0da9d5c56b92f4a0c524c12abe91" gracePeriod=15 Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.153412 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://d4fb86989c8a3e44185969c0a59b6e39def5e539bf58cfd51a7f305ffb37ae26" gracePeriod=15 Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.164882 4815 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.165945 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.165962 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.165971 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.165979 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.165992 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.165998 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.166011 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.166018 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.166111 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.166120 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.166128 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.166133 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.166143 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.166150 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.166169 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.166175 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.166339 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.166351 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.166359 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.166374 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.166387 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.166400 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.166411 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.166588 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.166595 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.166606 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.166612 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.166767 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.167089 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.168635 4815 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.169316 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.176661 4815 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.183769 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.183804 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4r69\" (UniqueName: \"kubernetes.io/projected/40e8b948-118a-404c-8683-3a800e159f03-kube-api-access-x4r69\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.183817 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnvcl\" (UniqueName: \"kubernetes.io/projected/9d9c770e-abbd-4c20-a96f-bf95d2a89d28-kube-api-access-lnvcl\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.183825 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40e8b948-118a-404c-8683-3a800e159f03-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.198134 4815 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.142:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.285031 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.285120 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8b5h\" (UniqueName: \"kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h\") pod \"auto-csr-approver-29533764-9mvkb\" (UID: \"ba102d89-25ea-40ad-ae0b-78746dbf6d2a\") " pod="openshift-infra/auto-csr-approver-29533764-9mvkb" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.285160 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.285188 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.285257 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.285374 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.285419 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.285528 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.285596 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.387230 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.387293 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8b5h\" (UniqueName: \"kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h\") pod \"auto-csr-approver-29533764-9mvkb\" (UID: \"ba102d89-25ea-40ad-ae0b-78746dbf6d2a\") " pod="openshift-infra/auto-csr-approver-29533764-9mvkb" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.387329 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.387357 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.387388 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.387411 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.387428 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.387460 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.387487 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.387612 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.387655 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.388022 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.388057 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.388085 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.388113 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.388141 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.388166 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.388274 4815 projected.go:194] Error preparing data for projected volume kube-api-access-k8b5h for pod openshift-infra/auto-csr-approver-29533764-9mvkb: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/serviceaccounts/csr-approver-sa/token": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.388321 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h podName:ba102d89-25ea-40ad-ae0b-78746dbf6d2a nodeName:}" failed. No retries permitted until 2026-02-25 13:24:00.888304156 +0000 UTC m=+198.689402220 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-k8b5h" (UniqueName: "kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h") pod "auto-csr-approver-29533764-9mvkb" (UID: "ba102d89-25ea-40ad-ae0b-78746dbf6d2a") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/serviceaccounts/csr-approver-sa/token": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.388602 4815 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/events\": dial tcp 38.102.83.142:6443: connect: connection refused" event="&Event{ObjectMeta:{auto-csr-approver-29533764-9mvkb.1897801cfd285818 openshift-infra 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-infra,Name:auto-csr-approver-29533764-9mvkb,UID:ba102d89-25ea-40ad-ae0b-78746dbf6d2a,APIVersion:v1,ResourceVersion:29897,FieldPath:,},Reason:FailedMount,Message:MountVolume.SetUp failed for volume \"kube-api-access-k8b5h\" : failed to fetch token: Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/serviceaccounts/csr-approver-sa/token\": dial tcp 38.102.83.142:6443: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:24:00.388298776 +0000 UTC m=+198.189396840,LastTimestamp:2026-02-25 13:24:00.388298776 +0000 UTC m=+198.189396840,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.499352 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: W0225 13:24:00.530363 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-80f67fe95608572e89bd410f3e8618d290a47996128ec29b2cf9a612eb8403a8 WatchSource:0}: Error finding container 80f67fe95608572e89bd410f3e8618d290a47996128ec29b2cf9a612eb8403a8: Status 404 returned error can't find the container with id 80f67fe95608572e89bd410f3e8618d290a47996128ec29b2cf9a612eb8403a8 Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.893395 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8b5h\" (UniqueName: \"kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h\") pod \"auto-csr-approver-29533764-9mvkb\" (UID: \"ba102d89-25ea-40ad-ae0b-78746dbf6d2a\") " pod="openshift-infra/auto-csr-approver-29533764-9mvkb" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.895008 4815 projected.go:194] Error preparing data for projected volume kube-api-access-k8b5h for pod openshift-infra/auto-csr-approver-29533764-9mvkb: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/serviceaccounts/csr-approver-sa/token": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.895126 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h podName:ba102d89-25ea-40ad-ae0b-78746dbf6d2a nodeName:}" failed. No retries permitted until 2026-02-25 13:24:01.895092606 +0000 UTC m=+199.696190710 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-k8b5h" (UniqueName: "kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h") pod "auto-csr-approver-29533764-9mvkb" (UID: "ba102d89-25ea-40ad-ae0b-78746dbf6d2a") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/serviceaccounts/csr-approver-sa/token": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.913473 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.915862 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.917289 4815 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="75f8f6e705409d40c4115c32c4e36f96c90c0da9d5c56b92f4a0c524c12abe91" exitCode=0 Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.917326 4815 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f25f11a9ab520446a86fd17fe4c1ffc9bf0622c1bbf72019dbb1535caba77279" exitCode=0 Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.917343 4815 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="38d1ca633d56103d9a2620de025cf43520585a65f1f225944a7a2f02714c8600" exitCode=0 Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.917362 4815 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fa8220d4ac2c87c277e7e2a77c5d619bfb5af22a635106397a1c4104d420a70c" exitCode=2 Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.917374 4815 scope.go:117] "RemoveContainer" containerID="0bb59977532b467793fb68d57745c41d814b3da95fcaef8a163ec0c6cc34ce02" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.920142 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"cfc2d669b8104d8e7d3ba7d6a28b62d6f2043fb943970ca6d5f531e85abeab5f"} Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.920229 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"80f67fe95608572e89bd410f3e8618d290a47996128ec29b2cf9a612eb8403a8"} Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.921248 4815 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: E0225 13:24:00.921295 4815 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.142:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.922786 4815 generic.go:334] "Generic (PLEG): container finished" podID="276236cd-c94d-4d4b-87e6-e333a1602548" containerID="de057a7e5b17a30c00200fb6cedfdb6137cd5db6b3d2919dda0c71f2b988c80b" exitCode=0 Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.922838 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"276236cd-c94d-4d4b-87e6-e333a1602548","Type":"ContainerDied","Data":"de057a7e5b17a30c00200fb6cedfdb6137cd5db6b3d2919dda0c71f2b988c80b"} Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.923673 4815 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.924214 4815 status_manager.go:851] "Failed to get status for pod" podUID="276236cd-c94d-4d4b-87e6-e333a1602548" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.927425 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.927436 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.927425 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" event={"ID":"9d9c770e-abbd-4c20-a96f-bf95d2a89d28","Type":"ContainerDied","Data":"259cada00403507ac09e8fd220104ddf3f0720762595a91a9a4ad629667b46ed"} Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.929599 4815 status_manager.go:851] "Failed to get status for pod" podUID="276236cd-c94d-4d4b-87e6-e333a1602548" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.929899 4815 status_manager.go:851] "Failed to get status for pod" podUID="40e8b948-118a-404c-8683-3a800e159f03" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-54767cd7bb-vdcdg\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.930338 4815 status_manager.go:851] "Failed to get status for pod" podUID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-74f875b898-j2txv\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.930748 4815 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.931271 4815 status_manager.go:851] "Failed to get status for pod" podUID="276236cd-c94d-4d4b-87e6-e333a1602548" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.931741 4815 status_manager.go:851] "Failed to get status for pod" podUID="40e8b948-118a-404c-8683-3a800e159f03" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-54767cd7bb-vdcdg\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.932124 4815 status_manager.go:851] "Failed to get status for pod" podUID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-74f875b898-j2txv\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.932503 4815 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.962057 4815 status_manager.go:851] "Failed to get status for pod" podUID="276236cd-c94d-4d4b-87e6-e333a1602548" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.962436 4815 status_manager.go:851] "Failed to get status for pod" podUID="40e8b948-118a-404c-8683-3a800e159f03" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-54767cd7bb-vdcdg\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.963260 4815 status_manager.go:851] "Failed to get status for pod" podUID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-74f875b898-j2txv\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.963830 4815 status_manager.go:851] "Failed to get status for pod" podUID="276236cd-c94d-4d4b-87e6-e333a1602548" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.964066 4815 status_manager.go:851] "Failed to get status for pod" podUID="40e8b948-118a-404c-8683-3a800e159f03" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-54767cd7bb-vdcdg\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.964230 4815 status_manager.go:851] "Failed to get status for pod" podUID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-74f875b898-j2txv\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:00 crc kubenswrapper[4815]: I0225 13:24:00.965964 4815 scope.go:117] "RemoveContainer" containerID="d9596a73a7255f1be66892c03c316906f3fa30994be69afc49392561e3465cea" Feb 25 13:24:01 crc kubenswrapper[4815]: I0225 13:24:01.908977 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8b5h\" (UniqueName: \"kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h\") pod \"auto-csr-approver-29533764-9mvkb\" (UID: \"ba102d89-25ea-40ad-ae0b-78746dbf6d2a\") " pod="openshift-infra/auto-csr-approver-29533764-9mvkb" Feb 25 13:24:01 crc kubenswrapper[4815]: E0225 13:24:01.909965 4815 projected.go:194] Error preparing data for projected volume kube-api-access-k8b5h for pod openshift-infra/auto-csr-approver-29533764-9mvkb: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/serviceaccounts/csr-approver-sa/token": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:24:01 crc kubenswrapper[4815]: E0225 13:24:01.910089 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h podName:ba102d89-25ea-40ad-ae0b-78746dbf6d2a nodeName:}" failed. No retries permitted until 2026-02-25 13:24:03.910042579 +0000 UTC m=+201.711140674 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-k8b5h" (UniqueName: "kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h") pod "auto-csr-approver-29533764-9mvkb" (UID: "ba102d89-25ea-40ad-ae0b-78746dbf6d2a") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/serviceaccounts/csr-approver-sa/token": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:24:01 crc kubenswrapper[4815]: I0225 13:24:01.942083 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.336884 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.337991 4815 status_manager.go:851] "Failed to get status for pod" podUID="276236cd-c94d-4d4b-87e6-e333a1602548" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.338453 4815 status_manager.go:851] "Failed to get status for pod" podUID="40e8b948-118a-404c-8683-3a800e159f03" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-54767cd7bb-vdcdg\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.338779 4815 status_manager.go:851] "Failed to get status for pod" podUID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-74f875b898-j2txv\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.418401 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/276236cd-c94d-4d4b-87e6-e333a1602548-kubelet-dir\") pod \"276236cd-c94d-4d4b-87e6-e333a1602548\" (UID: \"276236cd-c94d-4d4b-87e6-e333a1602548\") " Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.418464 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/276236cd-c94d-4d4b-87e6-e333a1602548-var-lock\") pod \"276236cd-c94d-4d4b-87e6-e333a1602548\" (UID: \"276236cd-c94d-4d4b-87e6-e333a1602548\") " Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.418586 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/276236cd-c94d-4d4b-87e6-e333a1602548-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "276236cd-c94d-4d4b-87e6-e333a1602548" (UID: "276236cd-c94d-4d4b-87e6-e333a1602548"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.418624 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/276236cd-c94d-4d4b-87e6-e333a1602548-var-lock" (OuterVolumeSpecName: "var-lock") pod "276236cd-c94d-4d4b-87e6-e333a1602548" (UID: "276236cd-c94d-4d4b-87e6-e333a1602548"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.418666 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/276236cd-c94d-4d4b-87e6-e333a1602548-kube-api-access\") pod \"276236cd-c94d-4d4b-87e6-e333a1602548\" (UID: \"276236cd-c94d-4d4b-87e6-e333a1602548\") " Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.418931 4815 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/276236cd-c94d-4d4b-87e6-e333a1602548-var-lock\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.418953 4815 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/276236cd-c94d-4d4b-87e6-e333a1602548-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.423660 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/276236cd-c94d-4d4b-87e6-e333a1602548-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "276236cd-c94d-4d4b-87e6-e333a1602548" (UID: "276236cd-c94d-4d4b-87e6-e333a1602548"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:24:02 crc kubenswrapper[4815]: E0225 13:24:02.425042 4815 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/events\": dial tcp 38.102.83.142:6443: connect: connection refused" event="&Event{ObjectMeta:{auto-csr-approver-29533764-9mvkb.1897801cfd285818 openshift-infra 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-infra,Name:auto-csr-approver-29533764-9mvkb,UID:ba102d89-25ea-40ad-ae0b-78746dbf6d2a,APIVersion:v1,ResourceVersion:29897,FieldPath:,},Reason:FailedMount,Message:MountVolume.SetUp failed for volume \"kube-api-access-k8b5h\" : failed to fetch token: Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/serviceaccounts/csr-approver-sa/token\": dial tcp 38.102.83.142:6443: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:24:00.388298776 +0000 UTC m=+198.189396840,LastTimestamp:2026-02-25 13:24:00.388298776 +0000 UTC m=+198.189396840,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.458543 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.459186 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.459669 4815 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.460013 4815 status_manager.go:851] "Failed to get status for pod" podUID="276236cd-c94d-4d4b-87e6-e333a1602548" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.460488 4815 status_manager.go:851] "Failed to get status for pod" podUID="40e8b948-118a-404c-8683-3a800e159f03" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-54767cd7bb-vdcdg\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.460842 4815 status_manager.go:851] "Failed to get status for pod" podUID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-74f875b898-j2txv\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.520651 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/276236cd-c94d-4d4b-87e6-e333a1602548-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.621268 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.621383 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.621422 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.621447 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.621456 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.621574 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.621774 4815 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.621792 4815 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.621806 4815 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.939420 4815 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.940268 4815 status_manager.go:851] "Failed to get status for pod" podUID="276236cd-c94d-4d4b-87e6-e333a1602548" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.941188 4815 status_manager.go:851] "Failed to get status for pod" podUID="40e8b948-118a-404c-8683-3a800e159f03" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-54767cd7bb-vdcdg\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.941386 4815 status_manager.go:851] "Failed to get status for pod" podUID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-74f875b898-j2txv\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.942276 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.952349 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.953020 4815 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d4fb86989c8a3e44185969c0a59b6e39def5e539bf58cfd51a7f305ffb37ae26" exitCode=0 Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.953125 4815 scope.go:117] "RemoveContainer" containerID="75f8f6e705409d40c4115c32c4e36f96c90c0da9d5c56b92f4a0c524c12abe91" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.953227 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.954446 4815 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.954947 4815 status_manager.go:851] "Failed to get status for pod" podUID="276236cd-c94d-4d4b-87e6-e333a1602548" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.955615 4815 status_manager.go:851] "Failed to get status for pod" podUID="40e8b948-118a-404c-8683-3a800e159f03" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-54767cd7bb-vdcdg\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.955912 4815 status_manager.go:851] "Failed to get status for pod" podUID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-74f875b898-j2txv\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.956028 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"276236cd-c94d-4d4b-87e6-e333a1602548","Type":"ContainerDied","Data":"d9b2f34bd19e97f149bc776d740d686d329196d8ee7f5d7ad452db22d7f75a77"} Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.956076 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9b2f34bd19e97f149bc776d740d686d329196d8ee7f5d7ad452db22d7f75a77" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.956080 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.970610 4815 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.971010 4815 status_manager.go:851] "Failed to get status for pod" podUID="276236cd-c94d-4d4b-87e6-e333a1602548" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.971256 4815 status_manager.go:851] "Failed to get status for pod" podUID="40e8b948-118a-404c-8683-3a800e159f03" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-54767cd7bb-vdcdg\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.971486 4815 status_manager.go:851] "Failed to get status for pod" podUID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-74f875b898-j2txv\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.975185 4815 status_manager.go:851] "Failed to get status for pod" podUID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-74f875b898-j2txv\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.977142 4815 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.978191 4815 status_manager.go:851] "Failed to get status for pod" podUID="276236cd-c94d-4d4b-87e6-e333a1602548" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.978611 4815 status_manager.go:851] "Failed to get status for pod" podUID="40e8b948-118a-404c-8683-3a800e159f03" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-54767cd7bb-vdcdg\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:02 crc kubenswrapper[4815]: I0225 13:24:02.986489 4815 scope.go:117] "RemoveContainer" containerID="f25f11a9ab520446a86fd17fe4c1ffc9bf0622c1bbf72019dbb1535caba77279" Feb 25 13:24:03 crc kubenswrapper[4815]: I0225 13:24:03.000143 4815 scope.go:117] "RemoveContainer" containerID="38d1ca633d56103d9a2620de025cf43520585a65f1f225944a7a2f02714c8600" Feb 25 13:24:03 crc kubenswrapper[4815]: I0225 13:24:03.015014 4815 scope.go:117] "RemoveContainer" containerID="fa8220d4ac2c87c277e7e2a77c5d619bfb5af22a635106397a1c4104d420a70c" Feb 25 13:24:03 crc kubenswrapper[4815]: I0225 13:24:03.030492 4815 scope.go:117] "RemoveContainer" containerID="d4fb86989c8a3e44185969c0a59b6e39def5e539bf58cfd51a7f305ffb37ae26" Feb 25 13:24:03 crc kubenswrapper[4815]: I0225 13:24:03.048566 4815 scope.go:117] "RemoveContainer" containerID="19311d205c069764f7d2d3ed470cf30dc0e47564b75ee0e51bc0fdcc5756965e" Feb 25 13:24:03 crc kubenswrapper[4815]: I0225 13:24:03.080662 4815 scope.go:117] "RemoveContainer" containerID="75f8f6e705409d40c4115c32c4e36f96c90c0da9d5c56b92f4a0c524c12abe91" Feb 25 13:24:03 crc kubenswrapper[4815]: E0225 13:24:03.081435 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75f8f6e705409d40c4115c32c4e36f96c90c0da9d5c56b92f4a0c524c12abe91\": container with ID starting with 75f8f6e705409d40c4115c32c4e36f96c90c0da9d5c56b92f4a0c524c12abe91 not found: ID does not exist" containerID="75f8f6e705409d40c4115c32c4e36f96c90c0da9d5c56b92f4a0c524c12abe91" Feb 25 13:24:03 crc kubenswrapper[4815]: I0225 13:24:03.081489 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75f8f6e705409d40c4115c32c4e36f96c90c0da9d5c56b92f4a0c524c12abe91"} err="failed to get container status \"75f8f6e705409d40c4115c32c4e36f96c90c0da9d5c56b92f4a0c524c12abe91\": rpc error: code = NotFound desc = could not find container \"75f8f6e705409d40c4115c32c4e36f96c90c0da9d5c56b92f4a0c524c12abe91\": container with ID starting with 75f8f6e705409d40c4115c32c4e36f96c90c0da9d5c56b92f4a0c524c12abe91 not found: ID does not exist" Feb 25 13:24:03 crc kubenswrapper[4815]: I0225 13:24:03.081540 4815 scope.go:117] "RemoveContainer" containerID="f25f11a9ab520446a86fd17fe4c1ffc9bf0622c1bbf72019dbb1535caba77279" Feb 25 13:24:03 crc kubenswrapper[4815]: E0225 13:24:03.082004 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f25f11a9ab520446a86fd17fe4c1ffc9bf0622c1bbf72019dbb1535caba77279\": container with ID starting with f25f11a9ab520446a86fd17fe4c1ffc9bf0622c1bbf72019dbb1535caba77279 not found: ID does not exist" containerID="f25f11a9ab520446a86fd17fe4c1ffc9bf0622c1bbf72019dbb1535caba77279" Feb 25 13:24:03 crc kubenswrapper[4815]: I0225 13:24:03.082053 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f25f11a9ab520446a86fd17fe4c1ffc9bf0622c1bbf72019dbb1535caba77279"} err="failed to get container status \"f25f11a9ab520446a86fd17fe4c1ffc9bf0622c1bbf72019dbb1535caba77279\": rpc error: code = NotFound desc = could not find container \"f25f11a9ab520446a86fd17fe4c1ffc9bf0622c1bbf72019dbb1535caba77279\": container with ID starting with f25f11a9ab520446a86fd17fe4c1ffc9bf0622c1bbf72019dbb1535caba77279 not found: ID does not exist" Feb 25 13:24:03 crc kubenswrapper[4815]: I0225 13:24:03.082073 4815 scope.go:117] "RemoveContainer" containerID="38d1ca633d56103d9a2620de025cf43520585a65f1f225944a7a2f02714c8600" Feb 25 13:24:03 crc kubenswrapper[4815]: E0225 13:24:03.082459 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38d1ca633d56103d9a2620de025cf43520585a65f1f225944a7a2f02714c8600\": container with ID starting with 38d1ca633d56103d9a2620de025cf43520585a65f1f225944a7a2f02714c8600 not found: ID does not exist" containerID="38d1ca633d56103d9a2620de025cf43520585a65f1f225944a7a2f02714c8600" Feb 25 13:24:03 crc kubenswrapper[4815]: I0225 13:24:03.082490 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38d1ca633d56103d9a2620de025cf43520585a65f1f225944a7a2f02714c8600"} err="failed to get container status \"38d1ca633d56103d9a2620de025cf43520585a65f1f225944a7a2f02714c8600\": rpc error: code = NotFound desc = could not find container \"38d1ca633d56103d9a2620de025cf43520585a65f1f225944a7a2f02714c8600\": container with ID starting with 38d1ca633d56103d9a2620de025cf43520585a65f1f225944a7a2f02714c8600 not found: ID does not exist" Feb 25 13:24:03 crc kubenswrapper[4815]: I0225 13:24:03.082523 4815 scope.go:117] "RemoveContainer" containerID="fa8220d4ac2c87c277e7e2a77c5d619bfb5af22a635106397a1c4104d420a70c" Feb 25 13:24:03 crc kubenswrapper[4815]: E0225 13:24:03.084965 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa8220d4ac2c87c277e7e2a77c5d619bfb5af22a635106397a1c4104d420a70c\": container with ID starting with fa8220d4ac2c87c277e7e2a77c5d619bfb5af22a635106397a1c4104d420a70c not found: ID does not exist" containerID="fa8220d4ac2c87c277e7e2a77c5d619bfb5af22a635106397a1c4104d420a70c" Feb 25 13:24:03 crc kubenswrapper[4815]: I0225 13:24:03.085002 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa8220d4ac2c87c277e7e2a77c5d619bfb5af22a635106397a1c4104d420a70c"} err="failed to get container status \"fa8220d4ac2c87c277e7e2a77c5d619bfb5af22a635106397a1c4104d420a70c\": rpc error: code = NotFound desc = could not find container \"fa8220d4ac2c87c277e7e2a77c5d619bfb5af22a635106397a1c4104d420a70c\": container with ID starting with fa8220d4ac2c87c277e7e2a77c5d619bfb5af22a635106397a1c4104d420a70c not found: ID does not exist" Feb 25 13:24:03 crc kubenswrapper[4815]: I0225 13:24:03.085055 4815 scope.go:117] "RemoveContainer" containerID="d4fb86989c8a3e44185969c0a59b6e39def5e539bf58cfd51a7f305ffb37ae26" Feb 25 13:24:03 crc kubenswrapper[4815]: E0225 13:24:03.085535 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4fb86989c8a3e44185969c0a59b6e39def5e539bf58cfd51a7f305ffb37ae26\": container with ID starting with d4fb86989c8a3e44185969c0a59b6e39def5e539bf58cfd51a7f305ffb37ae26 not found: ID does not exist" containerID="d4fb86989c8a3e44185969c0a59b6e39def5e539bf58cfd51a7f305ffb37ae26" Feb 25 13:24:03 crc kubenswrapper[4815]: I0225 13:24:03.085571 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4fb86989c8a3e44185969c0a59b6e39def5e539bf58cfd51a7f305ffb37ae26"} err="failed to get container status \"d4fb86989c8a3e44185969c0a59b6e39def5e539bf58cfd51a7f305ffb37ae26\": rpc error: code = NotFound desc = could not find container \"d4fb86989c8a3e44185969c0a59b6e39def5e539bf58cfd51a7f305ffb37ae26\": container with ID starting with d4fb86989c8a3e44185969c0a59b6e39def5e539bf58cfd51a7f305ffb37ae26 not found: ID does not exist" Feb 25 13:24:03 crc kubenswrapper[4815]: I0225 13:24:03.085591 4815 scope.go:117] "RemoveContainer" containerID="19311d205c069764f7d2d3ed470cf30dc0e47564b75ee0e51bc0fdcc5756965e" Feb 25 13:24:03 crc kubenswrapper[4815]: E0225 13:24:03.085966 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19311d205c069764f7d2d3ed470cf30dc0e47564b75ee0e51bc0fdcc5756965e\": container with ID starting with 19311d205c069764f7d2d3ed470cf30dc0e47564b75ee0e51bc0fdcc5756965e not found: ID does not exist" containerID="19311d205c069764f7d2d3ed470cf30dc0e47564b75ee0e51bc0fdcc5756965e" Feb 25 13:24:03 crc kubenswrapper[4815]: I0225 13:24:03.085994 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19311d205c069764f7d2d3ed470cf30dc0e47564b75ee0e51bc0fdcc5756965e"} err="failed to get container status \"19311d205c069764f7d2d3ed470cf30dc0e47564b75ee0e51bc0fdcc5756965e\": rpc error: code = NotFound desc = could not find container \"19311d205c069764f7d2d3ed470cf30dc0e47564b75ee0e51bc0fdcc5756965e\": container with ID starting with 19311d205c069764f7d2d3ed470cf30dc0e47564b75ee0e51bc0fdcc5756965e not found: ID does not exist" Feb 25 13:24:03 crc kubenswrapper[4815]: I0225 13:24:03.939713 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8b5h\" (UniqueName: \"kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h\") pod \"auto-csr-approver-29533764-9mvkb\" (UID: \"ba102d89-25ea-40ad-ae0b-78746dbf6d2a\") " pod="openshift-infra/auto-csr-approver-29533764-9mvkb" Feb 25 13:24:03 crc kubenswrapper[4815]: E0225 13:24:03.940613 4815 projected.go:194] Error preparing data for projected volume kube-api-access-k8b5h for pod openshift-infra/auto-csr-approver-29533764-9mvkb: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/serviceaccounts/csr-approver-sa/token": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:24:03 crc kubenswrapper[4815]: E0225 13:24:03.940713 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h podName:ba102d89-25ea-40ad-ae0b-78746dbf6d2a nodeName:}" failed. No retries permitted until 2026-02-25 13:24:07.940687761 +0000 UTC m=+205.741785865 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-k8b5h" (UniqueName: "kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h") pod "auto-csr-approver-29533764-9mvkb" (UID: "ba102d89-25ea-40ad-ae0b-78746dbf6d2a") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/serviceaccounts/csr-approver-sa/token": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:24:04 crc kubenswrapper[4815]: E0225 13:24:04.769027 4815 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:04 crc kubenswrapper[4815]: E0225 13:24:04.770199 4815 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:04 crc kubenswrapper[4815]: E0225 13:24:04.770971 4815 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:04 crc kubenswrapper[4815]: E0225 13:24:04.771313 4815 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:04 crc kubenswrapper[4815]: E0225 13:24:04.771827 4815 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:04 crc kubenswrapper[4815]: I0225 13:24:04.771937 4815 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 25 13:24:04 crc kubenswrapper[4815]: E0225 13:24:04.772314 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" interval="200ms" Feb 25 13:24:04 crc kubenswrapper[4815]: E0225 13:24:04.912967 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:24:04Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:24:04Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:24:04Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T13:24:04Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:04 crc kubenswrapper[4815]: E0225 13:24:04.913447 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:04 crc kubenswrapper[4815]: E0225 13:24:04.913769 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:04 crc kubenswrapper[4815]: E0225 13:24:04.914087 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:04 crc kubenswrapper[4815]: E0225 13:24:04.914435 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:04 crc kubenswrapper[4815]: E0225 13:24:04.914485 4815 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 25 13:24:04 crc kubenswrapper[4815]: E0225 13:24:04.972893 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" interval="400ms" Feb 25 13:24:05 crc kubenswrapper[4815]: E0225 13:24:05.373856 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" interval="800ms" Feb 25 13:24:06 crc kubenswrapper[4815]: E0225 13:24:06.176124 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" interval="1.6s" Feb 25 13:24:07 crc kubenswrapper[4815]: E0225 13:24:07.777491 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" interval="3.2s" Feb 25 13:24:07 crc kubenswrapper[4815]: I0225 13:24:07.997300 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8b5h\" (UniqueName: \"kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h\") pod \"auto-csr-approver-29533764-9mvkb\" (UID: \"ba102d89-25ea-40ad-ae0b-78746dbf6d2a\") " pod="openshift-infra/auto-csr-approver-29533764-9mvkb" Feb 25 13:24:07 crc kubenswrapper[4815]: E0225 13:24:07.998327 4815 projected.go:194] Error preparing data for projected volume kube-api-access-k8b5h for pod openshift-infra/auto-csr-approver-29533764-9mvkb: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/serviceaccounts/csr-approver-sa/token": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:24:07 crc kubenswrapper[4815]: E0225 13:24:07.998405 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h podName:ba102d89-25ea-40ad-ae0b-78746dbf6d2a nodeName:}" failed. No retries permitted until 2026-02-25 13:24:15.998383886 +0000 UTC m=+213.799481940 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-k8b5h" (UniqueName: "kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h") pod "auto-csr-approver-29533764-9mvkb" (UID: "ba102d89-25ea-40ad-ae0b-78746dbf6d2a") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/serviceaccounts/csr-approver-sa/token": dial tcp 38.102.83.142:6443: connect: connection refused Feb 25 13:24:08 crc kubenswrapper[4815]: E0225 13:24:08.940156 4815 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.142:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" volumeName="registry-storage" Feb 25 13:24:10 crc kubenswrapper[4815]: I0225 13:24:10.935018 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:10 crc kubenswrapper[4815]: I0225 13:24:10.936747 4815 status_manager.go:851] "Failed to get status for pod" podUID="276236cd-c94d-4d4b-87e6-e333a1602548" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:10 crc kubenswrapper[4815]: I0225 13:24:10.937651 4815 status_manager.go:851] "Failed to get status for pod" podUID="40e8b948-118a-404c-8683-3a800e159f03" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-54767cd7bb-vdcdg\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:10 crc kubenswrapper[4815]: I0225 13:24:10.938176 4815 status_manager.go:851] "Failed to get status for pod" podUID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-74f875b898-j2txv\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:10 crc kubenswrapper[4815]: I0225 13:24:10.960145 4815 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5168397a-19d2-4d16-af28-43ca3ceed269" Feb 25 13:24:10 crc kubenswrapper[4815]: I0225 13:24:10.960196 4815 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5168397a-19d2-4d16-af28-43ca3ceed269" Feb 25 13:24:10 crc kubenswrapper[4815]: E0225 13:24:10.960985 4815 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:10 crc kubenswrapper[4815]: I0225 13:24:10.961616 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:10 crc kubenswrapper[4815]: E0225 13:24:10.978722 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.142:6443: connect: connection refused" interval="6.4s" Feb 25 13:24:11 crc kubenswrapper[4815]: I0225 13:24:11.006669 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"73217a8f3a7f6e14a596d5abe68e54679030e155db909d925cbd1de4c88f57df"} Feb 25 13:24:12 crc kubenswrapper[4815]: I0225 13:24:12.017146 4815 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="225c29506d3e767e22139ef7354fae7f2e8d9767f537234489ea5cc9aa73a32e" exitCode=0 Feb 25 13:24:12 crc kubenswrapper[4815]: I0225 13:24:12.017215 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"225c29506d3e767e22139ef7354fae7f2e8d9767f537234489ea5cc9aa73a32e"} Feb 25 13:24:12 crc kubenswrapper[4815]: I0225 13:24:12.017602 4815 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5168397a-19d2-4d16-af28-43ca3ceed269" Feb 25 13:24:12 crc kubenswrapper[4815]: I0225 13:24:12.017641 4815 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5168397a-19d2-4d16-af28-43ca3ceed269" Feb 25 13:24:12 crc kubenswrapper[4815]: I0225 13:24:12.018148 4815 status_manager.go:851] "Failed to get status for pod" podUID="276236cd-c94d-4d4b-87e6-e333a1602548" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:12 crc kubenswrapper[4815]: E0225 13:24:12.018271 4815 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.142:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:12 crc kubenswrapper[4815]: I0225 13:24:12.018676 4815 status_manager.go:851] "Failed to get status for pod" podUID="40e8b948-118a-404c-8683-3a800e159f03" pod="openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-54767cd7bb-vdcdg\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:12 crc kubenswrapper[4815]: I0225 13:24:12.019339 4815 status_manager.go:851] "Failed to get status for pod" podUID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" pod="openshift-controller-manager/controller-manager-74f875b898-j2txv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-74f875b898-j2txv\": dial tcp 38.102.83.142:6443: connect: connection refused" Feb 25 13:24:12 crc kubenswrapper[4815]: E0225 13:24:12.426322 4815 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/events\": dial tcp 38.102.83.142:6443: connect: connection refused" event="&Event{ObjectMeta:{auto-csr-approver-29533764-9mvkb.1897801cfd285818 openshift-infra 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-infra,Name:auto-csr-approver-29533764-9mvkb,UID:ba102d89-25ea-40ad-ae0b-78746dbf6d2a,APIVersion:v1,ResourceVersion:29897,FieldPath:,},Reason:FailedMount,Message:MountVolume.SetUp failed for volume \"kube-api-access-k8b5h\" : failed to fetch token: Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/serviceaccounts/csr-approver-sa/token\": dial tcp 38.102.83.142:6443: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 13:24:00.388298776 +0000 UTC m=+198.189396840,LastTimestamp:2026-02-25 13:24:00.388298776 +0000 UTC m=+198.189396840,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 13:24:13 crc kubenswrapper[4815]: I0225 13:24:13.026837 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fa4a87fa6ac5d4886823545591a8597facb287b55fb79bc97dd28f1507e7204a"} Feb 25 13:24:13 crc kubenswrapper[4815]: I0225 13:24:13.026902 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"26c5a0241dadc37f63587364d2fe5cfc50322a1d0307134679a4d2d54510ee34"} Feb 25 13:24:13 crc kubenswrapper[4815]: I0225 13:24:13.026919 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e9c07f98b87015c709310c5758995abd6edb40b87184ad8e91758a7744756e08"} Feb 25 13:24:14 crc kubenswrapper[4815]: I0225 13:24:14.035335 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 25 13:24:14 crc kubenswrapper[4815]: I0225 13:24:14.036356 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 25 13:24:14 crc kubenswrapper[4815]: I0225 13:24:14.036399 4815 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="5d0d690f29a96ea3e6b07d3e6e5da8f22972ff02b52434520a7da40b5d1fa4ee" exitCode=1 Feb 25 13:24:14 crc kubenswrapper[4815]: I0225 13:24:14.036451 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"5d0d690f29a96ea3e6b07d3e6e5da8f22972ff02b52434520a7da40b5d1fa4ee"} Feb 25 13:24:14 crc kubenswrapper[4815]: I0225 13:24:14.036868 4815 scope.go:117] "RemoveContainer" containerID="5d0d690f29a96ea3e6b07d3e6e5da8f22972ff02b52434520a7da40b5d1fa4ee" Feb 25 13:24:14 crc kubenswrapper[4815]: I0225 13:24:14.039972 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"70c5568253ad5ea3ab7174e3578162100777d417049fdf1a0654bfc73d0a9ce8"} Feb 25 13:24:14 crc kubenswrapper[4815]: I0225 13:24:14.040013 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e10a7367a6a04fad10d3068a462613ed66868f71e697a939468c2b6205075407"} Feb 25 13:24:14 crc kubenswrapper[4815]: I0225 13:24:14.040187 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:14 crc kubenswrapper[4815]: I0225 13:24:14.040343 4815 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5168397a-19d2-4d16-af28-43ca3ceed269" Feb 25 13:24:14 crc kubenswrapper[4815]: I0225 13:24:14.040379 4815 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5168397a-19d2-4d16-af28-43ca3ceed269" Feb 25 13:24:14 crc kubenswrapper[4815]: I0225 13:24:14.723692 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:24:15 crc kubenswrapper[4815]: I0225 13:24:15.049768 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 25 13:24:15 crc kubenswrapper[4815]: I0225 13:24:15.051078 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 25 13:24:15 crc kubenswrapper[4815]: I0225 13:24:15.051246 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3f12bf9a7999ebed4ec8afb47d3e871b7becf78fd9c6d8fdec103f6951980d3e"} Feb 25 13:24:15 crc kubenswrapper[4815]: I0225 13:24:15.962582 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:15 crc kubenswrapper[4815]: I0225 13:24:15.962950 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:15 crc kubenswrapper[4815]: I0225 13:24:15.971329 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:16 crc kubenswrapper[4815]: I0225 13:24:16.042065 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8b5h\" (UniqueName: \"kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h\") pod \"auto-csr-approver-29533764-9mvkb\" (UID: \"ba102d89-25ea-40ad-ae0b-78746dbf6d2a\") " pod="openshift-infra/auto-csr-approver-29533764-9mvkb" Feb 25 13:24:16 crc kubenswrapper[4815]: I0225 13:24:16.078978 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8b5h\" (UniqueName: \"kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h\") pod \"auto-csr-approver-29533764-9mvkb\" (UID: \"ba102d89-25ea-40ad-ae0b-78746dbf6d2a\") " pod="openshift-infra/auto-csr-approver-29533764-9mvkb" Feb 25 13:24:16 crc kubenswrapper[4815]: I0225 13:24:16.348006 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533764-9mvkb" Feb 25 13:24:17 crc kubenswrapper[4815]: I0225 13:24:17.071748 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533764-9mvkb" event={"ID":"ba102d89-25ea-40ad-ae0b-78746dbf6d2a","Type":"ContainerStarted","Data":"bfa178d7fbea8882fee9661801d4822f00e58c1696de3fabdee46df311d9db6f"} Feb 25 13:24:19 crc kubenswrapper[4815]: I0225 13:24:19.088077 4815 generic.go:334] "Generic (PLEG): container finished" podID="ba102d89-25ea-40ad-ae0b-78746dbf6d2a" containerID="ef223c7f15d02f5311304de3f4d229a0c93eba794dc87c41fa03c34388bf23c1" exitCode=0 Feb 25 13:24:19 crc kubenswrapper[4815]: I0225 13:24:19.088143 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533764-9mvkb" event={"ID":"ba102d89-25ea-40ad-ae0b-78746dbf6d2a","Type":"ContainerDied","Data":"ef223c7f15d02f5311304de3f4d229a0c93eba794dc87c41fa03c34388bf23c1"} Feb 25 13:24:19 crc kubenswrapper[4815]: I0225 13:24:19.123630 4815 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:19 crc kubenswrapper[4815]: I0225 13:24:19.176585 4815 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="3a53ba81-4f3b-4609-a4e6-a61566c8d17f" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.041675 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" podUID="887d2231-1877-4977-ac7d-9f7cf1b33a23" containerName="oauth-openshift" containerID="cri-o://9ef793a39cc3472a5cd88bcb5c1ee71ede84f400225e179b7567a195dd768bf3" gracePeriod=15 Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.096162 4815 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5168397a-19d2-4d16-af28-43ca3ceed269" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.096246 4815 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5168397a-19d2-4d16-af28-43ca3ceed269" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.099110 4815 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="3a53ba81-4f3b-4609-a4e6-a61566c8d17f" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.402091 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533764-9mvkb" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.471980 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.500979 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8b5h\" (UniqueName: \"kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h\") pod \"ba102d89-25ea-40ad-ae0b-78746dbf6d2a\" (UID: \"ba102d89-25ea-40ad-ae0b-78746dbf6d2a\") " Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.507812 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h" (OuterVolumeSpecName: "kube-api-access-k8b5h") pod "ba102d89-25ea-40ad-ae0b-78746dbf6d2a" (UID: "ba102d89-25ea-40ad-ae0b-78746dbf6d2a"). InnerVolumeSpecName "kube-api-access-k8b5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.602905 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-cliconfig\") pod \"887d2231-1877-4977-ac7d-9f7cf1b33a23\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.603379 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/887d2231-1877-4977-ac7d-9f7cf1b33a23-audit-dir\") pod \"887d2231-1877-4977-ac7d-9f7cf1b33a23\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.603435 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-audit-policies\") pod \"887d2231-1877-4977-ac7d-9f7cf1b33a23\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.603492 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-trusted-ca-bundle\") pod \"887d2231-1877-4977-ac7d-9f7cf1b33a23\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.603597 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-ocp-branding-template\") pod \"887d2231-1877-4977-ac7d-9f7cf1b33a23\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.603635 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/887d2231-1877-4977-ac7d-9f7cf1b33a23-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "887d2231-1877-4977-ac7d-9f7cf1b33a23" (UID: "887d2231-1877-4977-ac7d-9f7cf1b33a23"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.603674 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86tlk\" (UniqueName: \"kubernetes.io/projected/887d2231-1877-4977-ac7d-9f7cf1b33a23-kube-api-access-86tlk\") pod \"887d2231-1877-4977-ac7d-9f7cf1b33a23\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.603819 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-idp-0-file-data\") pod \"887d2231-1877-4977-ac7d-9f7cf1b33a23\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.604240 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "887d2231-1877-4977-ac7d-9f7cf1b33a23" (UID: "887d2231-1877-4977-ac7d-9f7cf1b33a23"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.604394 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "887d2231-1877-4977-ac7d-9f7cf1b33a23" (UID: "887d2231-1877-4977-ac7d-9f7cf1b33a23"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.604618 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-error\") pod \"887d2231-1877-4977-ac7d-9f7cf1b33a23\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.604660 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "887d2231-1877-4977-ac7d-9f7cf1b33a23" (UID: "887d2231-1877-4977-ac7d-9f7cf1b33a23"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.604824 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-serving-cert\") pod \"887d2231-1877-4977-ac7d-9f7cf1b33a23\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.604903 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-login\") pod \"887d2231-1877-4977-ac7d-9f7cf1b33a23\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.604992 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-router-certs\") pod \"887d2231-1877-4977-ac7d-9f7cf1b33a23\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.605266 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-service-ca\") pod \"887d2231-1877-4977-ac7d-9f7cf1b33a23\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.605462 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-session\") pod \"887d2231-1877-4977-ac7d-9f7cf1b33a23\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.605682 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-provider-selection\") pod \"887d2231-1877-4977-ac7d-9f7cf1b33a23\" (UID: \"887d2231-1877-4977-ac7d-9f7cf1b33a23\") " Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.606191 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.606236 4815 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/887d2231-1877-4977-ac7d-9f7cf1b33a23-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.606263 4815 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.606328 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.606356 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8b5h\" (UniqueName: \"kubernetes.io/projected/ba102d89-25ea-40ad-ae0b-78746dbf6d2a-kube-api-access-k8b5h\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.607652 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "887d2231-1877-4977-ac7d-9f7cf1b33a23" (UID: "887d2231-1877-4977-ac7d-9f7cf1b33a23"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.608156 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/887d2231-1877-4977-ac7d-9f7cf1b33a23-kube-api-access-86tlk" (OuterVolumeSpecName: "kube-api-access-86tlk") pod "887d2231-1877-4977-ac7d-9f7cf1b33a23" (UID: "887d2231-1877-4977-ac7d-9f7cf1b33a23"). InnerVolumeSpecName "kube-api-access-86tlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.608527 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "887d2231-1877-4977-ac7d-9f7cf1b33a23" (UID: "887d2231-1877-4977-ac7d-9f7cf1b33a23"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.608940 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "887d2231-1877-4977-ac7d-9f7cf1b33a23" (UID: "887d2231-1877-4977-ac7d-9f7cf1b33a23"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.609849 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "887d2231-1877-4977-ac7d-9f7cf1b33a23" (UID: "887d2231-1877-4977-ac7d-9f7cf1b33a23"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.610440 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "887d2231-1877-4977-ac7d-9f7cf1b33a23" (UID: "887d2231-1877-4977-ac7d-9f7cf1b33a23"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.610704 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "887d2231-1877-4977-ac7d-9f7cf1b33a23" (UID: "887d2231-1877-4977-ac7d-9f7cf1b33a23"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.611701 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "887d2231-1877-4977-ac7d-9f7cf1b33a23" (UID: "887d2231-1877-4977-ac7d-9f7cf1b33a23"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.613859 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "887d2231-1877-4977-ac7d-9f7cf1b33a23" (UID: "887d2231-1877-4977-ac7d-9f7cf1b33a23"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.614222 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "887d2231-1877-4977-ac7d-9f7cf1b33a23" (UID: "887d2231-1877-4977-ac7d-9f7cf1b33a23"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.708232 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86tlk\" (UniqueName: \"kubernetes.io/projected/887d2231-1877-4977-ac7d-9f7cf1b33a23-kube-api-access-86tlk\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.708280 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.708304 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.708327 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.708346 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.708367 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.708387 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.708556 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.708587 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:20 crc kubenswrapper[4815]: I0225 13:24:20.708651 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/887d2231-1877-4977-ac7d-9f7cf1b33a23-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:21 crc kubenswrapper[4815]: I0225 13:24:21.103645 4815 generic.go:334] "Generic (PLEG): container finished" podID="887d2231-1877-4977-ac7d-9f7cf1b33a23" containerID="9ef793a39cc3472a5cd88bcb5c1ee71ede84f400225e179b7567a195dd768bf3" exitCode=0 Feb 25 13:24:21 crc kubenswrapper[4815]: I0225 13:24:21.103767 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" event={"ID":"887d2231-1877-4977-ac7d-9f7cf1b33a23","Type":"ContainerDied","Data":"9ef793a39cc3472a5cd88bcb5c1ee71ede84f400225e179b7567a195dd768bf3"} Feb 25 13:24:21 crc kubenswrapper[4815]: I0225 13:24:21.103858 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" event={"ID":"887d2231-1877-4977-ac7d-9f7cf1b33a23","Type":"ContainerDied","Data":"e9e01707c67d715db60c7a341dfa3799191bef99c3a9bc52eb77adaa3b2f4e67"} Feb 25 13:24:21 crc kubenswrapper[4815]: I0225 13:24:21.103890 4815 scope.go:117] "RemoveContainer" containerID="9ef793a39cc3472a5cd88bcb5c1ee71ede84f400225e179b7567a195dd768bf3" Feb 25 13:24:21 crc kubenswrapper[4815]: I0225 13:24:21.103792 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-md7bn" Feb 25 13:24:21 crc kubenswrapper[4815]: I0225 13:24:21.112372 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533764-9mvkb" event={"ID":"ba102d89-25ea-40ad-ae0b-78746dbf6d2a","Type":"ContainerDied","Data":"bfa178d7fbea8882fee9661801d4822f00e58c1696de3fabdee46df311d9db6f"} Feb 25 13:24:21 crc kubenswrapper[4815]: I0225 13:24:21.112489 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfa178d7fbea8882fee9661801d4822f00e58c1696de3fabdee46df311d9db6f" Feb 25 13:24:21 crc kubenswrapper[4815]: I0225 13:24:21.113579 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533764-9mvkb" Feb 25 13:24:21 crc kubenswrapper[4815]: I0225 13:24:21.138367 4815 scope.go:117] "RemoveContainer" containerID="9ef793a39cc3472a5cd88bcb5c1ee71ede84f400225e179b7567a195dd768bf3" Feb 25 13:24:21 crc kubenswrapper[4815]: E0225 13:24:21.138833 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ef793a39cc3472a5cd88bcb5c1ee71ede84f400225e179b7567a195dd768bf3\": container with ID starting with 9ef793a39cc3472a5cd88bcb5c1ee71ede84f400225e179b7567a195dd768bf3 not found: ID does not exist" containerID="9ef793a39cc3472a5cd88bcb5c1ee71ede84f400225e179b7567a195dd768bf3" Feb 25 13:24:21 crc kubenswrapper[4815]: I0225 13:24:21.138888 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ef793a39cc3472a5cd88bcb5c1ee71ede84f400225e179b7567a195dd768bf3"} err="failed to get container status \"9ef793a39cc3472a5cd88bcb5c1ee71ede84f400225e179b7567a195dd768bf3\": rpc error: code = NotFound desc = could not find container \"9ef793a39cc3472a5cd88bcb5c1ee71ede84f400225e179b7567a195dd768bf3\": container with ID starting with 9ef793a39cc3472a5cd88bcb5c1ee71ede84f400225e179b7567a195dd768bf3 not found: ID does not exist" Feb 25 13:24:21 crc kubenswrapper[4815]: I0225 13:24:21.342187 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:24:21 crc kubenswrapper[4815]: I0225 13:24:21.342254 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:24:22 crc kubenswrapper[4815]: I0225 13:24:22.372595 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:24:23 crc kubenswrapper[4815]: I0225 13:24:23.299140 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:24:23 crc kubenswrapper[4815]: I0225 13:24:23.299314 4815 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Feb 25 13:24:23 crc kubenswrapper[4815]: I0225 13:24:23.299356 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Feb 25 13:24:27 crc kubenswrapper[4815]: I0225 13:24:27.195242 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 25 13:24:27 crc kubenswrapper[4815]: I0225 13:24:27.252642 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 25 13:24:27 crc kubenswrapper[4815]: I0225 13:24:27.417338 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 25 13:24:28 crc kubenswrapper[4815]: I0225 13:24:28.345660 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 25 13:24:29 crc kubenswrapper[4815]: I0225 13:24:29.179225 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 25 13:24:29 crc kubenswrapper[4815]: I0225 13:24:29.410004 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 25 13:24:30 crc kubenswrapper[4815]: I0225 13:24:30.081146 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 25 13:24:30 crc kubenswrapper[4815]: I0225 13:24:30.127692 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 25 13:24:30 crc kubenswrapper[4815]: I0225 13:24:30.605975 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 25 13:24:30 crc kubenswrapper[4815]: I0225 13:24:30.673454 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 25 13:24:30 crc kubenswrapper[4815]: I0225 13:24:30.926500 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 25 13:24:30 crc kubenswrapper[4815]: I0225 13:24:30.986111 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 25 13:24:31 crc kubenswrapper[4815]: I0225 13:24:31.022037 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 25 13:24:31 crc kubenswrapper[4815]: I0225 13:24:31.299367 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 25 13:24:31 crc kubenswrapper[4815]: I0225 13:24:31.675964 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 25 13:24:31 crc kubenswrapper[4815]: I0225 13:24:31.900871 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 25 13:24:31 crc kubenswrapper[4815]: I0225 13:24:31.947737 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 25 13:24:32 crc kubenswrapper[4815]: I0225 13:24:32.173854 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 25 13:24:32 crc kubenswrapper[4815]: I0225 13:24:32.814770 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 25 13:24:32 crc kubenswrapper[4815]: I0225 13:24:32.914416 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.005082 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.025030 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.063140 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.122241 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.154761 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.242351 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.252497 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.299818 4815 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.299893 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.571609 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.571841 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.609751 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.675201 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.864200 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.901036 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.906269 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.913981 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.925912 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.987454 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 25 13:24:33 crc kubenswrapper[4815]: I0225 13:24:33.992098 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.083461 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.192164 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.386703 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.479107 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.524104 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.547677 4815 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.554934 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-74f875b898-j2txv","openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-md7bn","openshift-route-controller-manager/route-controller-manager-54767cd7bb-vdcdg"] Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.555053 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-76654bf648-xfvz4","openshift-kube-apiserver/kube-apiserver-crc","openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg"] Feb 25 13:24:34 crc kubenswrapper[4815]: E0225 13:24:34.555384 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="887d2231-1877-4977-ac7d-9f7cf1b33a23" containerName="oauth-openshift" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.555428 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="887d2231-1877-4977-ac7d-9f7cf1b33a23" containerName="oauth-openshift" Feb 25 13:24:34 crc kubenswrapper[4815]: E0225 13:24:34.555459 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="276236cd-c94d-4d4b-87e6-e333a1602548" containerName="installer" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.555478 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="276236cd-c94d-4d4b-87e6-e333a1602548" containerName="installer" Feb 25 13:24:34 crc kubenswrapper[4815]: E0225 13:24:34.555501 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba102d89-25ea-40ad-ae0b-78746dbf6d2a" containerName="oc" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.555552 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba102d89-25ea-40ad-ae0b-78746dbf6d2a" containerName="oc" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.555773 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="887d2231-1877-4977-ac7d-9f7cf1b33a23" containerName="oauth-openshift" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.555787 4815 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5168397a-19d2-4d16-af28-43ca3ceed269" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.555828 4815 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5168397a-19d2-4d16-af28-43ca3ceed269" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.555803 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba102d89-25ea-40ad-ae0b-78746dbf6d2a" containerName="oc" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.556032 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="276236cd-c94d-4d4b-87e6-e333a1602548" containerName="installer" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.556645 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.556688 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533764-9mvkb"] Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.556740 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.561247 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.561293 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.561490 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.562619 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.563192 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.563257 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.563711 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.564265 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.564725 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.565229 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.565458 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.565630 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.565980 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.567727 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.572455 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.582150 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=15.582130627 podStartE2EDuration="15.582130627s" podCreationTimestamp="2026-02-25 13:24:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:24:34.580903136 +0000 UTC m=+232.382001230" watchObservedRunningTime="2026-02-25 13:24:34.582130627 +0000 UTC m=+232.383228681" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.704379 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q7pm\" (UniqueName: \"kubernetes.io/projected/d537a11a-fe15-4cfd-9dbb-35d7cc72edb9-kube-api-access-6q7pm\") pod \"route-controller-manager-99bcc65bd-5fnfg\" (UID: \"d537a11a-fe15-4cfd-9dbb-35d7cc72edb9\") " pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.704449 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d537a11a-fe15-4cfd-9dbb-35d7cc72edb9-config\") pod \"route-controller-manager-99bcc65bd-5fnfg\" (UID: \"d537a11a-fe15-4cfd-9dbb-35d7cc72edb9\") " pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.704638 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d537a11a-fe15-4cfd-9dbb-35d7cc72edb9-serving-cert\") pod \"route-controller-manager-99bcc65bd-5fnfg\" (UID: \"d537a11a-fe15-4cfd-9dbb-35d7cc72edb9\") " pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.704684 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18847355-96d2-4aac-b2e0-7689d1c3c477-serving-cert\") pod \"controller-manager-76654bf648-xfvz4\" (UID: \"18847355-96d2-4aac-b2e0-7689d1c3c477\") " pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.704721 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtrrx\" (UniqueName: \"kubernetes.io/projected/18847355-96d2-4aac-b2e0-7689d1c3c477-kube-api-access-xtrrx\") pod \"controller-manager-76654bf648-xfvz4\" (UID: \"18847355-96d2-4aac-b2e0-7689d1c3c477\") " pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.704760 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d537a11a-fe15-4cfd-9dbb-35d7cc72edb9-client-ca\") pod \"route-controller-manager-99bcc65bd-5fnfg\" (UID: \"d537a11a-fe15-4cfd-9dbb-35d7cc72edb9\") " pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.704808 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18847355-96d2-4aac-b2e0-7689d1c3c477-config\") pod \"controller-manager-76654bf648-xfvz4\" (UID: \"18847355-96d2-4aac-b2e0-7689d1c3c477\") " pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.704856 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18847355-96d2-4aac-b2e0-7689d1c3c477-client-ca\") pod \"controller-manager-76654bf648-xfvz4\" (UID: \"18847355-96d2-4aac-b2e0-7689d1c3c477\") " pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.704892 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/18847355-96d2-4aac-b2e0-7689d1c3c477-proxy-ca-bundles\") pod \"controller-manager-76654bf648-xfvz4\" (UID: \"18847355-96d2-4aac-b2e0-7689d1c3c477\") " pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.805694 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18847355-96d2-4aac-b2e0-7689d1c3c477-client-ca\") pod \"controller-manager-76654bf648-xfvz4\" (UID: \"18847355-96d2-4aac-b2e0-7689d1c3c477\") " pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.805759 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/18847355-96d2-4aac-b2e0-7689d1c3c477-proxy-ca-bundles\") pod \"controller-manager-76654bf648-xfvz4\" (UID: \"18847355-96d2-4aac-b2e0-7689d1c3c477\") " pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.805799 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q7pm\" (UniqueName: \"kubernetes.io/projected/d537a11a-fe15-4cfd-9dbb-35d7cc72edb9-kube-api-access-6q7pm\") pod \"route-controller-manager-99bcc65bd-5fnfg\" (UID: \"d537a11a-fe15-4cfd-9dbb-35d7cc72edb9\") " pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.805835 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d537a11a-fe15-4cfd-9dbb-35d7cc72edb9-config\") pod \"route-controller-manager-99bcc65bd-5fnfg\" (UID: \"d537a11a-fe15-4cfd-9dbb-35d7cc72edb9\") " pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.805917 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d537a11a-fe15-4cfd-9dbb-35d7cc72edb9-serving-cert\") pod \"route-controller-manager-99bcc65bd-5fnfg\" (UID: \"d537a11a-fe15-4cfd-9dbb-35d7cc72edb9\") " pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.805942 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18847355-96d2-4aac-b2e0-7689d1c3c477-serving-cert\") pod \"controller-manager-76654bf648-xfvz4\" (UID: \"18847355-96d2-4aac-b2e0-7689d1c3c477\") " pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.805964 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtrrx\" (UniqueName: \"kubernetes.io/projected/18847355-96d2-4aac-b2e0-7689d1c3c477-kube-api-access-xtrrx\") pod \"controller-manager-76654bf648-xfvz4\" (UID: \"18847355-96d2-4aac-b2e0-7689d1c3c477\") " pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.805986 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d537a11a-fe15-4cfd-9dbb-35d7cc72edb9-client-ca\") pod \"route-controller-manager-99bcc65bd-5fnfg\" (UID: \"d537a11a-fe15-4cfd-9dbb-35d7cc72edb9\") " pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.806017 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18847355-96d2-4aac-b2e0-7689d1c3c477-config\") pod \"controller-manager-76654bf648-xfvz4\" (UID: \"18847355-96d2-4aac-b2e0-7689d1c3c477\") " pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.807444 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18847355-96d2-4aac-b2e0-7689d1c3c477-client-ca\") pod \"controller-manager-76654bf648-xfvz4\" (UID: \"18847355-96d2-4aac-b2e0-7689d1c3c477\") " pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.807901 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18847355-96d2-4aac-b2e0-7689d1c3c477-config\") pod \"controller-manager-76654bf648-xfvz4\" (UID: \"18847355-96d2-4aac-b2e0-7689d1c3c477\") " pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.808009 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d537a11a-fe15-4cfd-9dbb-35d7cc72edb9-config\") pod \"route-controller-manager-99bcc65bd-5fnfg\" (UID: \"d537a11a-fe15-4cfd-9dbb-35d7cc72edb9\") " pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.808126 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d537a11a-fe15-4cfd-9dbb-35d7cc72edb9-client-ca\") pod \"route-controller-manager-99bcc65bd-5fnfg\" (UID: \"d537a11a-fe15-4cfd-9dbb-35d7cc72edb9\") " pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.809196 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/18847355-96d2-4aac-b2e0-7689d1c3c477-proxy-ca-bundles\") pod \"controller-manager-76654bf648-xfvz4\" (UID: \"18847355-96d2-4aac-b2e0-7689d1c3c477\") " pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.813930 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d537a11a-fe15-4cfd-9dbb-35d7cc72edb9-serving-cert\") pod \"route-controller-manager-99bcc65bd-5fnfg\" (UID: \"d537a11a-fe15-4cfd-9dbb-35d7cc72edb9\") " pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.819430 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.821337 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18847355-96d2-4aac-b2e0-7689d1c3c477-serving-cert\") pod \"controller-manager-76654bf648-xfvz4\" (UID: \"18847355-96d2-4aac-b2e0-7689d1c3c477\") " pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.828631 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtrrx\" (UniqueName: \"kubernetes.io/projected/18847355-96d2-4aac-b2e0-7689d1c3c477-kube-api-access-xtrrx\") pod \"controller-manager-76654bf648-xfvz4\" (UID: \"18847355-96d2-4aac-b2e0-7689d1c3c477\") " pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.834563 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q7pm\" (UniqueName: \"kubernetes.io/projected/d537a11a-fe15-4cfd-9dbb-35d7cc72edb9-kube-api-access-6q7pm\") pod \"route-controller-manager-99bcc65bd-5fnfg\" (UID: \"d537a11a-fe15-4cfd-9dbb-35d7cc72edb9\") " pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.890641 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.899150 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.910071 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.949285 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40e8b948-118a-404c-8683-3a800e159f03" path="/var/lib/kubelet/pods/40e8b948-118a-404c-8683-3a800e159f03/volumes" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.951223 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="887d2231-1877-4977-ac7d-9f7cf1b33a23" path="/var/lib/kubelet/pods/887d2231-1877-4977-ac7d-9f7cf1b33a23/volumes" Feb 25 13:24:34 crc kubenswrapper[4815]: I0225 13:24:34.952463 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d9c770e-abbd-4c20-a96f-bf95d2a89d28" path="/var/lib/kubelet/pods/9d9c770e-abbd-4c20-a96f-bf95d2a89d28/volumes" Feb 25 13:24:35 crc kubenswrapper[4815]: I0225 13:24:35.168833 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 25 13:24:35 crc kubenswrapper[4815]: I0225 13:24:35.227885 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 25 13:24:35 crc kubenswrapper[4815]: I0225 13:24:35.235386 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 25 13:24:35 crc kubenswrapper[4815]: I0225 13:24:35.328393 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 25 13:24:35 crc kubenswrapper[4815]: I0225 13:24:35.414901 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 25 13:24:35 crc kubenswrapper[4815]: I0225 13:24:35.524332 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 25 13:24:35 crc kubenswrapper[4815]: I0225 13:24:35.681342 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 25 13:24:35 crc kubenswrapper[4815]: I0225 13:24:35.687978 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 25 13:24:35 crc kubenswrapper[4815]: I0225 13:24:35.690923 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 25 13:24:35 crc kubenswrapper[4815]: I0225 13:24:35.824812 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 25 13:24:35 crc kubenswrapper[4815]: I0225 13:24:35.826544 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 25 13:24:35 crc kubenswrapper[4815]: I0225 13:24:35.834354 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 25 13:24:35 crc kubenswrapper[4815]: I0225 13:24:35.922353 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 25 13:24:35 crc kubenswrapper[4815]: I0225 13:24:35.923548 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 25 13:24:36 crc kubenswrapper[4815]: I0225 13:24:36.033179 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 25 13:24:36 crc kubenswrapper[4815]: I0225 13:24:36.142689 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 25 13:24:36 crc kubenswrapper[4815]: I0225 13:24:36.150674 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 25 13:24:36 crc kubenswrapper[4815]: I0225 13:24:36.178173 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 25 13:24:36 crc kubenswrapper[4815]: I0225 13:24:36.184487 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 25 13:24:36 crc kubenswrapper[4815]: I0225 13:24:36.257933 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 25 13:24:36 crc kubenswrapper[4815]: I0225 13:24:36.434739 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 25 13:24:36 crc kubenswrapper[4815]: I0225 13:24:36.453478 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 25 13:24:36 crc kubenswrapper[4815]: I0225 13:24:36.471037 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 25 13:24:36 crc kubenswrapper[4815]: I0225 13:24:36.474425 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 25 13:24:36 crc kubenswrapper[4815]: I0225 13:24:36.503845 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 25 13:24:36 crc kubenswrapper[4815]: I0225 13:24:36.536975 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 25 13:24:36 crc kubenswrapper[4815]: I0225 13:24:36.557098 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 25 13:24:36 crc kubenswrapper[4815]: I0225 13:24:36.637930 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 25 13:24:36 crc kubenswrapper[4815]: I0225 13:24:36.744420 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 25 13:24:36 crc kubenswrapper[4815]: I0225 13:24:36.981728 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 25 13:24:36 crc kubenswrapper[4815]: I0225 13:24:36.981927 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 25 13:24:37 crc kubenswrapper[4815]: I0225 13:24:37.010066 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 25 13:24:37 crc kubenswrapper[4815]: I0225 13:24:37.185353 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 25 13:24:37 crc kubenswrapper[4815]: I0225 13:24:37.217197 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 25 13:24:37 crc kubenswrapper[4815]: I0225 13:24:37.326494 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 25 13:24:37 crc kubenswrapper[4815]: I0225 13:24:37.736670 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 25 13:24:37 crc kubenswrapper[4815]: I0225 13:24:37.759532 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 25 13:24:37 crc kubenswrapper[4815]: I0225 13:24:37.763303 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 25 13:24:37 crc kubenswrapper[4815]: I0225 13:24:37.772986 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 25 13:24:37 crc kubenswrapper[4815]: I0225 13:24:37.784882 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 25 13:24:37 crc kubenswrapper[4815]: I0225 13:24:37.863771 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 25 13:24:37 crc kubenswrapper[4815]: I0225 13:24:37.956664 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 25 13:24:37 crc kubenswrapper[4815]: I0225 13:24:37.962197 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 25 13:24:37 crc kubenswrapper[4815]: I0225 13:24:37.996019 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.152544 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.278047 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.302950 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.343705 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.455016 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-fc667b7f-btjv7"] Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.455672 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.458754 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.458948 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.459196 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.459313 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.459432 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.459580 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.459898 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.460153 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.460408 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.460815 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.461017 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.461408 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.469195 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.478554 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.479301 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.479953 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.549803 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-serving-cert\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.549868 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-user-template-error\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.549912 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-user-template-login\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.550036 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/32198d61-405e-4edb-b682-1fc869f64386-audit-dir\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.550077 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.550111 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-service-ca\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.550152 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.550183 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-session\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.550262 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.550328 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-router-certs\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.550363 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcjz4\" (UniqueName: \"kubernetes.io/projected/32198d61-405e-4edb-b682-1fc869f64386-kube-api-access-jcjz4\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.550406 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/32198d61-405e-4edb-b682-1fc869f64386-audit-policies\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.550441 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-cliconfig\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.550586 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.575092 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.651551 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.651624 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-serving-cert\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.651653 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-user-template-error\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.651679 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-user-template-login\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.651704 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/32198d61-405e-4edb-b682-1fc869f64386-audit-dir\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.651724 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.651745 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-service-ca\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.651763 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-session\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.651819 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/32198d61-405e-4edb-b682-1fc869f64386-audit-dir\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.651858 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.652459 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-service-ca\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.652523 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.652566 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-router-certs\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.652586 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcjz4\" (UniqueName: \"kubernetes.io/projected/32198d61-405e-4edb-b682-1fc869f64386-kube-api-access-jcjz4\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.652883 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/32198d61-405e-4edb-b682-1fc869f64386-audit-policies\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.652909 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-cliconfig\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.653175 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.653369 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-cliconfig\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.653716 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/32198d61-405e-4edb-b682-1fc869f64386-audit-policies\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.656900 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-router-certs\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.656908 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.656995 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-session\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.658404 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.661898 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-user-template-login\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.670726 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-system-serving-cert\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.664882 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.670961 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/32198d61-405e-4edb-b682-1fc869f64386-v4-0-config-user-template-error\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.674689 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcjz4\" (UniqueName: \"kubernetes.io/projected/32198d61-405e-4edb-b682-1fc869f64386-kube-api-access-jcjz4\") pod \"oauth-openshift-fc667b7f-btjv7\" (UID: \"32198d61-405e-4edb-b682-1fc869f64386\") " pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.755205 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.764925 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.770250 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.927159 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 25 13:24:38 crc kubenswrapper[4815]: I0225 13:24:38.976341 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.096731 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.114195 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.173864 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.229760 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.266975 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.287483 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.302218 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.345735 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.358937 4815 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.477588 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.498663 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.542611 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.572905 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.611750 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.638110 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.657499 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.736654 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.800611 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.930652 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 25 13:24:39 crc kubenswrapper[4815]: I0225 13:24:39.991066 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 25 13:24:40 crc kubenswrapper[4815]: I0225 13:24:40.008992 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 25 13:24:40 crc kubenswrapper[4815]: I0225 13:24:40.013322 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 25 13:24:40 crc kubenswrapper[4815]: I0225 13:24:40.064111 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 25 13:24:40 crc kubenswrapper[4815]: I0225 13:24:40.169411 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 25 13:24:40 crc kubenswrapper[4815]: I0225 13:24:40.190869 4815 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 25 13:24:40 crc kubenswrapper[4815]: I0225 13:24:40.336800 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 25 13:24:40 crc kubenswrapper[4815]: I0225 13:24:40.343163 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 25 13:24:40 crc kubenswrapper[4815]: I0225 13:24:40.457137 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 25 13:24:40 crc kubenswrapper[4815]: I0225 13:24:40.561196 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 25 13:24:40 crc kubenswrapper[4815]: I0225 13:24:40.572750 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 25 13:24:40 crc kubenswrapper[4815]: I0225 13:24:40.677087 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 25 13:24:40 crc kubenswrapper[4815]: I0225 13:24:40.740078 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 25 13:24:40 crc kubenswrapper[4815]: I0225 13:24:40.843800 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 25 13:24:40 crc kubenswrapper[4815]: I0225 13:24:40.886962 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 25 13:24:40 crc kubenswrapper[4815]: I0225 13:24:40.919245 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 25 13:24:40 crc kubenswrapper[4815]: I0225 13:24:40.938711 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.131810 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.208435 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.369048 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.432575 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.473504 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.588234 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.607035 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.660288 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.727806 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.772495 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.793644 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.837107 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.846217 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.860375 4815 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.860575 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://cfc2d669b8104d8e7d3ba7d6a28b62d6f2043fb943970ca6d5f531e85abeab5f" gracePeriod=5 Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.869475 4815 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.870214 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 25 13:24:41 crc kubenswrapper[4815]: I0225 13:24:41.967457 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 25 13:24:42 crc kubenswrapper[4815]: I0225 13:24:42.069049 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 25 13:24:42 crc kubenswrapper[4815]: I0225 13:24:42.078866 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 25 13:24:42 crc kubenswrapper[4815]: I0225 13:24:42.115221 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 25 13:24:42 crc kubenswrapper[4815]: I0225 13:24:42.118454 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 25 13:24:42 crc kubenswrapper[4815]: I0225 13:24:42.155229 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 25 13:24:42 crc kubenswrapper[4815]: I0225 13:24:42.207147 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 25 13:24:42 crc kubenswrapper[4815]: I0225 13:24:42.218565 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 25 13:24:42 crc kubenswrapper[4815]: I0225 13:24:42.353222 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 25 13:24:42 crc kubenswrapper[4815]: I0225 13:24:42.569154 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 25 13:24:42 crc kubenswrapper[4815]: I0225 13:24:42.606780 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 25 13:24:42 crc kubenswrapper[4815]: I0225 13:24:42.607070 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 25 13:24:42 crc kubenswrapper[4815]: I0225 13:24:42.703089 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 25 13:24:42 crc kubenswrapper[4815]: I0225 13:24:42.708321 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 25 13:24:42 crc kubenswrapper[4815]: I0225 13:24:42.712549 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 25 13:24:42 crc kubenswrapper[4815]: I0225 13:24:42.752619 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 25 13:24:42 crc kubenswrapper[4815]: I0225 13:24:42.864378 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 25 13:24:42 crc kubenswrapper[4815]: I0225 13:24:42.984229 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.068295 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.081647 4815 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.153367 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.197578 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.289884 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.299869 4815 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.300228 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.300502 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.301987 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"3f12bf9a7999ebed4ec8afb47d3e871b7becf78fd9c6d8fdec103f6951980d3e"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.302425 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://3f12bf9a7999ebed4ec8afb47d3e871b7becf78fd9c6d8fdec103f6951980d3e" gracePeriod=30 Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.324474 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.349563 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.426169 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.570006 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.615777 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.657336 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.732254 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.812141 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.828619 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.906993 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.912871 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 25 13:24:43 crc kubenswrapper[4815]: I0225 13:24:43.973447 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 25 13:24:44 crc kubenswrapper[4815]: I0225 13:24:44.195945 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 25 13:24:44 crc kubenswrapper[4815]: I0225 13:24:44.363812 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 25 13:24:44 crc kubenswrapper[4815]: I0225 13:24:44.449467 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 25 13:24:44 crc kubenswrapper[4815]: I0225 13:24:44.487268 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 25 13:24:44 crc kubenswrapper[4815]: I0225 13:24:44.529030 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 25 13:24:44 crc kubenswrapper[4815]: I0225 13:24:44.542475 4815 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 25 13:24:44 crc kubenswrapper[4815]: I0225 13:24:44.695416 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 25 13:24:44 crc kubenswrapper[4815]: I0225 13:24:44.748443 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 25 13:24:44 crc kubenswrapper[4815]: I0225 13:24:44.828991 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 25 13:24:44 crc kubenswrapper[4815]: I0225 13:24:44.838141 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 25 13:24:44 crc kubenswrapper[4815]: I0225 13:24:44.953167 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 25 13:24:44 crc kubenswrapper[4815]: I0225 13:24:44.964818 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 25 13:24:44 crc kubenswrapper[4815]: I0225 13:24:44.996942 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 25 13:24:45 crc kubenswrapper[4815]: I0225 13:24:45.187222 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 25 13:24:45 crc kubenswrapper[4815]: I0225 13:24:45.373102 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 25 13:24:45 crc kubenswrapper[4815]: I0225 13:24:45.390323 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 25 13:24:45 crc kubenswrapper[4815]: I0225 13:24:45.394317 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 25 13:24:45 crc kubenswrapper[4815]: I0225 13:24:45.429613 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 25 13:24:45 crc kubenswrapper[4815]: I0225 13:24:45.578568 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 25 13:24:45 crc kubenswrapper[4815]: I0225 13:24:45.687029 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 25 13:24:45 crc kubenswrapper[4815]: I0225 13:24:45.824254 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 25 13:24:45 crc kubenswrapper[4815]: I0225 13:24:45.865893 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 25 13:24:45 crc kubenswrapper[4815]: I0225 13:24:45.991480 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 25 13:24:46 crc kubenswrapper[4815]: I0225 13:24:46.025627 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 25 13:24:46 crc kubenswrapper[4815]: I0225 13:24:46.041982 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 25 13:24:46 crc kubenswrapper[4815]: I0225 13:24:46.196441 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 25 13:24:46 crc kubenswrapper[4815]: I0225 13:24:46.320304 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 25 13:24:46 crc kubenswrapper[4815]: I0225 13:24:46.433476 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 25 13:24:46 crc kubenswrapper[4815]: I0225 13:24:46.736604 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 25 13:24:46 crc kubenswrapper[4815]: I0225 13:24:46.767711 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 25 13:24:46 crc kubenswrapper[4815]: I0225 13:24:46.808304 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 25 13:24:47 crc kubenswrapper[4815]: E0225 13:24:47.005692 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-conmon-cfc2d669b8104d8e7d3ba7d6a28b62d6f2043fb943970ca6d5f531e85abeab5f.scope\": RecentStats: unable to find data in memory cache]" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.009243 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.118832 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.132940 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.228114 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.284711 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.285141 4815 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="cfc2d669b8104d8e7d3ba7d6a28b62d6f2043fb943970ca6d5f531e85abeab5f" exitCode=137 Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.460986 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.461063 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.572277 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.572348 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.572384 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.572430 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.572456 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.572483 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.572543 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.572636 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.572707 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.573335 4815 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.573365 4815 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.573377 4815 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.573386 4815 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.583728 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.676018 4815 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.733438 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.783434 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76654bf648-xfvz4"] Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.790946 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-fc667b7f-btjv7"] Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.802142 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.808650 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg"] Feb 25 13:24:47 crc kubenswrapper[4815]: I0225 13:24:47.950058 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-fc667b7f-btjv7"] Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.077208 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76654bf648-xfvz4"] Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.126773 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg"] Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.291773 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" event={"ID":"32198d61-405e-4edb-b682-1fc869f64386","Type":"ContainerStarted","Data":"a603a47aa7d556fa65ad6f89300c40d8fa05d4f3ad590316082dea7959288db3"} Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.291845 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.291857 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" event={"ID":"32198d61-405e-4edb-b682-1fc869f64386","Type":"ContainerStarted","Data":"41f91b73fb3f73295e3a1fe83c8d55087120597a9814416aae50e85c7a8eaddc"} Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.293339 4815 patch_prober.go:28] interesting pod/oauth-openshift-fc667b7f-btjv7 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.67:6443/healthz\": dial tcp 10.217.0.67:6443: connect: connection refused" start-of-body= Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.293408 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" podUID="32198d61-405e-4edb-b682-1fc869f64386" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.67:6443/healthz\": dial tcp 10.217.0.67:6443: connect: connection refused" Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.293569 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.293686 4815 scope.go:117] "RemoveContainer" containerID="cfc2d669b8104d8e7d3ba7d6a28b62d6f2043fb943970ca6d5f531e85abeab5f" Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.293697 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.296906 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" event={"ID":"d537a11a-fe15-4cfd-9dbb-35d7cc72edb9","Type":"ContainerStarted","Data":"4edf508649a19a491c3416a0f096b6d8693edc20fecd8891110a8212e860b897"} Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.296959 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" event={"ID":"d537a11a-fe15-4cfd-9dbb-35d7cc72edb9","Type":"ContainerStarted","Data":"ddfc2a64af5eb0236074288a9e896b6246e9406124a70fcec8137777686c829a"} Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.297086 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.298681 4815 patch_prober.go:28] interesting pod/route-controller-manager-99bcc65bd-5fnfg container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.69:8443/healthz\": dial tcp 10.217.0.69:8443: connect: connection refused" start-of-body= Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.298719 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" podUID="d537a11a-fe15-4cfd-9dbb-35d7cc72edb9" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.69:8443/healthz\": dial tcp 10.217.0.69:8443: connect: connection refused" Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.298805 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.298912 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" event={"ID":"18847355-96d2-4aac-b2e0-7689d1c3c477","Type":"ContainerStarted","Data":"4bde15c9cf9ee8915ecddef19629aa1e115b949ed6aeabbe5f870c8f7361eea0"} Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.298949 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" event={"ID":"18847355-96d2-4aac-b2e0-7689d1c3c477","Type":"ContainerStarted","Data":"45288dd2fe57480ce35cbe4703f8094728721b3206ef40b2119e3fcc2706cc93"} Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.299156 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.300337 4815 patch_prober.go:28] interesting pod/controller-manager-76654bf648-xfvz4 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" start-of-body= Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.300737 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" podUID="18847355-96d2-4aac-b2e0-7689d1c3c477" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.318734 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" podStartSLOduration=53.318699172 podStartE2EDuration="53.318699172s" podCreationTimestamp="2026-02-25 13:23:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:24:48.31498771 +0000 UTC m=+246.116085764" watchObservedRunningTime="2026-02-25 13:24:48.318699172 +0000 UTC m=+246.119797226" Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.339981 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" podStartSLOduration=49.339964218 podStartE2EDuration="49.339964218s" podCreationTimestamp="2026-02-25 13:23:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:24:48.337290331 +0000 UTC m=+246.138388385" watchObservedRunningTime="2026-02-25 13:24:48.339964218 +0000 UTC m=+246.141062262" Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.355808 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" podStartSLOduration=49.355793958 podStartE2EDuration="49.355793958s" podCreationTimestamp="2026-02-25 13:23:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:24:48.35282892 +0000 UTC m=+246.153926984" watchObservedRunningTime="2026-02-25 13:24:48.355793958 +0000 UTC m=+246.156892012" Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.775526 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-fc667b7f-btjv7" Feb 25 13:24:48 crc kubenswrapper[4815]: I0225 13:24:48.941750 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 25 13:24:49 crc kubenswrapper[4815]: I0225 13:24:49.314913 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-76654bf648-xfvz4" Feb 25 13:24:49 crc kubenswrapper[4815]: I0225 13:24:49.316301 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-99bcc65bd-5fnfg" Feb 25 13:24:51 crc kubenswrapper[4815]: I0225 13:24:51.342636 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:24:51 crc kubenswrapper[4815]: I0225 13:24:51.343233 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:25:13 crc kubenswrapper[4815]: I0225 13:25:13.474192 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Feb 25 13:25:13 crc kubenswrapper[4815]: I0225 13:25:13.476821 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 25 13:25:13 crc kubenswrapper[4815]: I0225 13:25:13.477526 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 25 13:25:13 crc kubenswrapper[4815]: I0225 13:25:13.477592 4815 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="3f12bf9a7999ebed4ec8afb47d3e871b7becf78fd9c6d8fdec103f6951980d3e" exitCode=137 Feb 25 13:25:13 crc kubenswrapper[4815]: I0225 13:25:13.477633 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"3f12bf9a7999ebed4ec8afb47d3e871b7becf78fd9c6d8fdec103f6951980d3e"} Feb 25 13:25:13 crc kubenswrapper[4815]: I0225 13:25:13.477678 4815 scope.go:117] "RemoveContainer" containerID="5d0d690f29a96ea3e6b07d3e6e5da8f22972ff02b52434520a7da40b5d1fa4ee" Feb 25 13:25:14 crc kubenswrapper[4815]: I0225 13:25:14.486262 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Feb 25 13:25:14 crc kubenswrapper[4815]: I0225 13:25:14.488555 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 25 13:25:14 crc kubenswrapper[4815]: I0225 13:25:14.488619 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1026a7345570a66ebb269d9ddf70dcd9af2a2baa61a466594ccacfee9f913b7d"} Feb 25 13:25:21 crc kubenswrapper[4815]: I0225 13:25:21.342206 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:25:21 crc kubenswrapper[4815]: I0225 13:25:21.342881 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:25:21 crc kubenswrapper[4815]: I0225 13:25:21.342966 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:25:21 crc kubenswrapper[4815]: I0225 13:25:21.343893 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"440b745816f3d3542a46a14793d9d030a841eaa34e91cd17885d1961ecc85bee"} pod="openshift-machine-config-operator/machine-config-daemon-frc75" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 13:25:21 crc kubenswrapper[4815]: I0225 13:25:21.343998 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" containerID="cri-o://440b745816f3d3542a46a14793d9d030a841eaa34e91cd17885d1961ecc85bee" gracePeriod=600 Feb 25 13:25:21 crc kubenswrapper[4815]: I0225 13:25:21.532003 4815 generic.go:334] "Generic (PLEG): container finished" podID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerID="440b745816f3d3542a46a14793d9d030a841eaa34e91cd17885d1961ecc85bee" exitCode=0 Feb 25 13:25:21 crc kubenswrapper[4815]: I0225 13:25:21.532046 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerDied","Data":"440b745816f3d3542a46a14793d9d030a841eaa34e91cd17885d1961ecc85bee"} Feb 25 13:25:22 crc kubenswrapper[4815]: I0225 13:25:22.371680 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:25:22 crc kubenswrapper[4815]: I0225 13:25:22.539783 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerStarted","Data":"0bde58aaffd292f31f5aeec24700d4368aa27e806f31f168fd069eb408b17058"} Feb 25 13:25:23 crc kubenswrapper[4815]: I0225 13:25:23.299185 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:25:23 crc kubenswrapper[4815]: I0225 13:25:23.304933 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:25:23 crc kubenswrapper[4815]: I0225 13:25:23.556978 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.427353 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vzljd"] Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.428521 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vzljd" podUID="c0706f0a-7f64-47d9-b3be-afe1d760603f" containerName="registry-server" containerID="cri-o://ba6d7bc00ccdc194c823eea650bd84480e9545de9ba1158a9d9711d07c69862f" gracePeriod=30 Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.440960 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zm6qq"] Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.441327 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zm6qq" podUID="9c7db856-98cb-47a4-bad2-35ec76191138" containerName="registry-server" containerID="cri-o://95e998a6ac1c1396f48b6482b190372c7309388aa8da448275af3ba3f6b9fd1a" gracePeriod=30 Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.453369 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6th6z"] Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.453717 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" podUID="1b2fb1a6-e1d1-48e0-83dc-17143971294e" containerName="marketplace-operator" containerID="cri-o://3ae1c93addc3743f6a3c4f03cb5cc5f88484ac039710085e2e9e90d700a80eac" gracePeriod=30 Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.457991 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6zz5"] Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.458313 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x6zz5" podUID="6d00f8b4-1bf6-4a47-9beb-2226f26d695b" containerName="registry-server" containerID="cri-o://4eeb3c6973f20f0a56118b5dbf28db1a25436109bf0347e578acad431a1051b9" gracePeriod=30 Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.468006 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9pjrm"] Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.468225 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9pjrm" podUID="bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" containerName="registry-server" containerID="cri-o://8248af535da5869c4f84a7e1453d2b519a4ff9c0ea2de89b618c2dcb5ea1040b" gracePeriod=30 Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.476118 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ltn9g"] Feb 25 13:25:53 crc kubenswrapper[4815]: E0225 13:25:53.476498 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.476536 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.476667 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.477185 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ltn9g" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.484834 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ltn9g"] Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.663064 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4685a3d5-0377-4fb7-a496-854c23dd3729-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ltn9g\" (UID: \"4685a3d5-0377-4fb7-a496-854c23dd3729\") " pod="openshift-marketplace/marketplace-operator-79b997595-ltn9g" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.663120 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4685a3d5-0377-4fb7-a496-854c23dd3729-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ltn9g\" (UID: \"4685a3d5-0377-4fb7-a496-854c23dd3729\") " pod="openshift-marketplace/marketplace-operator-79b997595-ltn9g" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.663148 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2kwz\" (UniqueName: \"kubernetes.io/projected/4685a3d5-0377-4fb7-a496-854c23dd3729-kube-api-access-x2kwz\") pod \"marketplace-operator-79b997595-ltn9g\" (UID: \"4685a3d5-0377-4fb7-a496-854c23dd3729\") " pod="openshift-marketplace/marketplace-operator-79b997595-ltn9g" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.718353 4815 generic.go:334] "Generic (PLEG): container finished" podID="1b2fb1a6-e1d1-48e0-83dc-17143971294e" containerID="3ae1c93addc3743f6a3c4f03cb5cc5f88484ac039710085e2e9e90d700a80eac" exitCode=0 Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.718433 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" event={"ID":"1b2fb1a6-e1d1-48e0-83dc-17143971294e","Type":"ContainerDied","Data":"3ae1c93addc3743f6a3c4f03cb5cc5f88484ac039710085e2e9e90d700a80eac"} Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.721446 4815 generic.go:334] "Generic (PLEG): container finished" podID="bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" containerID="8248af535da5869c4f84a7e1453d2b519a4ff9c0ea2de89b618c2dcb5ea1040b" exitCode=0 Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.721501 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9pjrm" event={"ID":"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4","Type":"ContainerDied","Data":"8248af535da5869c4f84a7e1453d2b519a4ff9c0ea2de89b618c2dcb5ea1040b"} Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.723713 4815 generic.go:334] "Generic (PLEG): container finished" podID="c0706f0a-7f64-47d9-b3be-afe1d760603f" containerID="ba6d7bc00ccdc194c823eea650bd84480e9545de9ba1158a9d9711d07c69862f" exitCode=0 Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.723739 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzljd" event={"ID":"c0706f0a-7f64-47d9-b3be-afe1d760603f","Type":"ContainerDied","Data":"ba6d7bc00ccdc194c823eea650bd84480e9545de9ba1158a9d9711d07c69862f"} Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.726780 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zm6qq" event={"ID":"9c7db856-98cb-47a4-bad2-35ec76191138","Type":"ContainerDied","Data":"95e998a6ac1c1396f48b6482b190372c7309388aa8da448275af3ba3f6b9fd1a"} Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.726669 4815 generic.go:334] "Generic (PLEG): container finished" podID="9c7db856-98cb-47a4-bad2-35ec76191138" containerID="95e998a6ac1c1396f48b6482b190372c7309388aa8da448275af3ba3f6b9fd1a" exitCode=0 Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.730079 4815 generic.go:334] "Generic (PLEG): container finished" podID="6d00f8b4-1bf6-4a47-9beb-2226f26d695b" containerID="4eeb3c6973f20f0a56118b5dbf28db1a25436109bf0347e578acad431a1051b9" exitCode=0 Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.730173 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6zz5" event={"ID":"6d00f8b4-1bf6-4a47-9beb-2226f26d695b","Type":"ContainerDied","Data":"4eeb3c6973f20f0a56118b5dbf28db1a25436109bf0347e578acad431a1051b9"} Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.764451 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4685a3d5-0377-4fb7-a496-854c23dd3729-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ltn9g\" (UID: \"4685a3d5-0377-4fb7-a496-854c23dd3729\") " pod="openshift-marketplace/marketplace-operator-79b997595-ltn9g" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.764499 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4685a3d5-0377-4fb7-a496-854c23dd3729-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ltn9g\" (UID: \"4685a3d5-0377-4fb7-a496-854c23dd3729\") " pod="openshift-marketplace/marketplace-operator-79b997595-ltn9g" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.764533 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2kwz\" (UniqueName: \"kubernetes.io/projected/4685a3d5-0377-4fb7-a496-854c23dd3729-kube-api-access-x2kwz\") pod \"marketplace-operator-79b997595-ltn9g\" (UID: \"4685a3d5-0377-4fb7-a496-854c23dd3729\") " pod="openshift-marketplace/marketplace-operator-79b997595-ltn9g" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.767876 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4685a3d5-0377-4fb7-a496-854c23dd3729-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ltn9g\" (UID: \"4685a3d5-0377-4fb7-a496-854c23dd3729\") " pod="openshift-marketplace/marketplace-operator-79b997595-ltn9g" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.779021 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4685a3d5-0377-4fb7-a496-854c23dd3729-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ltn9g\" (UID: \"4685a3d5-0377-4fb7-a496-854c23dd3729\") " pod="openshift-marketplace/marketplace-operator-79b997595-ltn9g" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.784184 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2kwz\" (UniqueName: \"kubernetes.io/projected/4685a3d5-0377-4fb7-a496-854c23dd3729-kube-api-access-x2kwz\") pod \"marketplace-operator-79b997595-ltn9g\" (UID: \"4685a3d5-0377-4fb7-a496-854c23dd3729\") " pod="openshift-marketplace/marketplace-operator-79b997595-ltn9g" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.797250 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ltn9g" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.910470 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vzljd" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.949679 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.971286 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9pjrm" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.974572 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6zz5" Feb 25 13:25:53 crc kubenswrapper[4815]: I0225 13:25:53.983465 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zm6qq" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.074969 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlqlg\" (UniqueName: \"kubernetes.io/projected/c0706f0a-7f64-47d9-b3be-afe1d760603f-kube-api-access-rlqlg\") pod \"c0706f0a-7f64-47d9-b3be-afe1d760603f\" (UID: \"c0706f0a-7f64-47d9-b3be-afe1d760603f\") " Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.075036 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cz6fn\" (UniqueName: \"kubernetes.io/projected/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-kube-api-access-cz6fn\") pod \"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4\" (UID: \"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4\") " Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.075071 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km5v7\" (UniqueName: \"kubernetes.io/projected/1b2fb1a6-e1d1-48e0-83dc-17143971294e-kube-api-access-km5v7\") pod \"1b2fb1a6-e1d1-48e0-83dc-17143971294e\" (UID: \"1b2fb1a6-e1d1-48e0-83dc-17143971294e\") " Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.075120 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zspjb\" (UniqueName: \"kubernetes.io/projected/9c7db856-98cb-47a4-bad2-35ec76191138-kube-api-access-zspjb\") pod \"9c7db856-98cb-47a4-bad2-35ec76191138\" (UID: \"9c7db856-98cb-47a4-bad2-35ec76191138\") " Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.075204 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-utilities\") pod \"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4\" (UID: \"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4\") " Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.075233 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-catalog-content\") pod \"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4\" (UID: \"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4\") " Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.075322 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0706f0a-7f64-47d9-b3be-afe1d760603f-utilities\") pod \"c0706f0a-7f64-47d9-b3be-afe1d760603f\" (UID: \"c0706f0a-7f64-47d9-b3be-afe1d760603f\") " Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.088752 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fw8ct\" (UniqueName: \"kubernetes.io/projected/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-kube-api-access-fw8ct\") pod \"6d00f8b4-1bf6-4a47-9beb-2226f26d695b\" (UID: \"6d00f8b4-1bf6-4a47-9beb-2226f26d695b\") " Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.088810 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c7db856-98cb-47a4-bad2-35ec76191138-utilities\") pod \"9c7db856-98cb-47a4-bad2-35ec76191138\" (UID: \"9c7db856-98cb-47a4-bad2-35ec76191138\") " Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.088846 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c7db856-98cb-47a4-bad2-35ec76191138-catalog-content\") pod \"9c7db856-98cb-47a4-bad2-35ec76191138\" (UID: \"9c7db856-98cb-47a4-bad2-35ec76191138\") " Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.088890 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0706f0a-7f64-47d9-b3be-afe1d760603f-catalog-content\") pod \"c0706f0a-7f64-47d9-b3be-afe1d760603f\" (UID: \"c0706f0a-7f64-47d9-b3be-afe1d760603f\") " Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.088942 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1b2fb1a6-e1d1-48e0-83dc-17143971294e-marketplace-operator-metrics\") pod \"1b2fb1a6-e1d1-48e0-83dc-17143971294e\" (UID: \"1b2fb1a6-e1d1-48e0-83dc-17143971294e\") " Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.088983 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b2fb1a6-e1d1-48e0-83dc-17143971294e-marketplace-trusted-ca\") pod \"1b2fb1a6-e1d1-48e0-83dc-17143971294e\" (UID: \"1b2fb1a6-e1d1-48e0-83dc-17143971294e\") " Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.089020 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-catalog-content\") pod \"6d00f8b4-1bf6-4a47-9beb-2226f26d695b\" (UID: \"6d00f8b4-1bf6-4a47-9beb-2226f26d695b\") " Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.076006 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-utilities" (OuterVolumeSpecName: "utilities") pod "bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" (UID: "bdf40b79-9414-4062-8dfd-3f0f71fcd1e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.076541 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0706f0a-7f64-47d9-b3be-afe1d760603f-utilities" (OuterVolumeSpecName: "utilities") pod "c0706f0a-7f64-47d9-b3be-afe1d760603f" (UID: "c0706f0a-7f64-47d9-b3be-afe1d760603f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.080070 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b2fb1a6-e1d1-48e0-83dc-17143971294e-kube-api-access-km5v7" (OuterVolumeSpecName: "kube-api-access-km5v7") pod "1b2fb1a6-e1d1-48e0-83dc-17143971294e" (UID: "1b2fb1a6-e1d1-48e0-83dc-17143971294e"). InnerVolumeSpecName "kube-api-access-km5v7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.082400 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-kube-api-access-cz6fn" (OuterVolumeSpecName: "kube-api-access-cz6fn") pod "bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" (UID: "bdf40b79-9414-4062-8dfd-3f0f71fcd1e4"). InnerVolumeSpecName "kube-api-access-cz6fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.082651 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c7db856-98cb-47a4-bad2-35ec76191138-kube-api-access-zspjb" (OuterVolumeSpecName: "kube-api-access-zspjb") pod "9c7db856-98cb-47a4-bad2-35ec76191138" (UID: "9c7db856-98cb-47a4-bad2-35ec76191138"). InnerVolumeSpecName "kube-api-access-zspjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.083281 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0706f0a-7f64-47d9-b3be-afe1d760603f-kube-api-access-rlqlg" (OuterVolumeSpecName: "kube-api-access-rlqlg") pod "c0706f0a-7f64-47d9-b3be-afe1d760603f" (UID: "c0706f0a-7f64-47d9-b3be-afe1d760603f"). InnerVolumeSpecName "kube-api-access-rlqlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.090468 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c7db856-98cb-47a4-bad2-35ec76191138-utilities" (OuterVolumeSpecName: "utilities") pod "9c7db856-98cb-47a4-bad2-35ec76191138" (UID: "9c7db856-98cb-47a4-bad2-35ec76191138"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.092227 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b2fb1a6-e1d1-48e0-83dc-17143971294e-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "1b2fb1a6-e1d1-48e0-83dc-17143971294e" (UID: "1b2fb1a6-e1d1-48e0-83dc-17143971294e"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.096794 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2fb1a6-e1d1-48e0-83dc-17143971294e-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "1b2fb1a6-e1d1-48e0-83dc-17143971294e" (UID: "1b2fb1a6-e1d1-48e0-83dc-17143971294e"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.096827 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-kube-api-access-fw8ct" (OuterVolumeSpecName: "kube-api-access-fw8ct") pod "6d00f8b4-1bf6-4a47-9beb-2226f26d695b" (UID: "6d00f8b4-1bf6-4a47-9beb-2226f26d695b"). InnerVolumeSpecName "kube-api-access-fw8ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.100742 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ltn9g"] Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.153187 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d00f8b4-1bf6-4a47-9beb-2226f26d695b" (UID: "6d00f8b4-1bf6-4a47-9beb-2226f26d695b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.155689 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c7db856-98cb-47a4-bad2-35ec76191138-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c7db856-98cb-47a4-bad2-35ec76191138" (UID: "9c7db856-98cb-47a4-bad2-35ec76191138"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.169518 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0706f0a-7f64-47d9-b3be-afe1d760603f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c0706f0a-7f64-47d9-b3be-afe1d760603f" (UID: "c0706f0a-7f64-47d9-b3be-afe1d760603f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.190367 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-utilities\") pod \"6d00f8b4-1bf6-4a47-9beb-2226f26d695b\" (UID: \"6d00f8b4-1bf6-4a47-9beb-2226f26d695b\") " Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.190718 4815 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b2fb1a6-e1d1-48e0-83dc-17143971294e-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.190746 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.190760 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cz6fn\" (UniqueName: \"kubernetes.io/projected/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-kube-api-access-cz6fn\") on node \"crc\" DevicePath \"\"" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.190774 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlqlg\" (UniqueName: \"kubernetes.io/projected/c0706f0a-7f64-47d9-b3be-afe1d760603f-kube-api-access-rlqlg\") on node \"crc\" DevicePath \"\"" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.190787 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km5v7\" (UniqueName: \"kubernetes.io/projected/1b2fb1a6-e1d1-48e0-83dc-17143971294e-kube-api-access-km5v7\") on node \"crc\" DevicePath \"\"" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.190799 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zspjb\" (UniqueName: \"kubernetes.io/projected/9c7db856-98cb-47a4-bad2-35ec76191138-kube-api-access-zspjb\") on node \"crc\" DevicePath \"\"" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.190810 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.190824 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0706f0a-7f64-47d9-b3be-afe1d760603f-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.190835 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fw8ct\" (UniqueName: \"kubernetes.io/projected/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-kube-api-access-fw8ct\") on node \"crc\" DevicePath \"\"" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.190845 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c7db856-98cb-47a4-bad2-35ec76191138-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.190858 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c7db856-98cb-47a4-bad2-35ec76191138-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.190870 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0706f0a-7f64-47d9-b3be-afe1d760603f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.190881 4815 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1b2fb1a6-e1d1-48e0-83dc-17143971294e-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.191030 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-utilities" (OuterVolumeSpecName: "utilities") pod "6d00f8b4-1bf6-4a47-9beb-2226f26d695b" (UID: "6d00f8b4-1bf6-4a47-9beb-2226f26d695b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.228680 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" (UID: "bdf40b79-9414-4062-8dfd-3f0f71fcd1e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.292549 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d00f8b4-1bf6-4a47-9beb-2226f26d695b-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.292590 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.736010 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zm6qq" event={"ID":"9c7db856-98cb-47a4-bad2-35ec76191138","Type":"ContainerDied","Data":"96628b7b501006f4311cedab3503ef1041488bc91b51ffa7b2d8727c9c43b7de"} Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.736053 4815 scope.go:117] "RemoveContainer" containerID="95e998a6ac1c1396f48b6482b190372c7309388aa8da448275af3ba3f6b9fd1a" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.736143 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zm6qq" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.753270 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6zz5" event={"ID":"6d00f8b4-1bf6-4a47-9beb-2226f26d695b","Type":"ContainerDied","Data":"0a5ffda5565243ea6d11416bdfcc0373a0a1457db88d02797cd1097a8be04ab7"} Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.753378 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6zz5" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.763143 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.763125 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6th6z" event={"ID":"1b2fb1a6-e1d1-48e0-83dc-17143971294e","Type":"ContainerDied","Data":"f447cce8a81086097513032f03607144a55ef1b9b23ec49f2a727d8a96b55a04"} Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.770424 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9pjrm" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.770967 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9pjrm" event={"ID":"bdf40b79-9414-4062-8dfd-3f0f71fcd1e4","Type":"ContainerDied","Data":"bdae89adda41be4ab717561875f8ddbcd505a0da76baa92e429fd3508d056cd4"} Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.774616 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ltn9g" event={"ID":"4685a3d5-0377-4fb7-a496-854c23dd3729","Type":"ContainerStarted","Data":"c7b9fd6ea9d0d96fd86116961500c8c6b1d17e657deb1a75af6eb6ac86e615a9"} Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.774658 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ltn9g" event={"ID":"4685a3d5-0377-4fb7-a496-854c23dd3729","Type":"ContainerStarted","Data":"5df84e1215783263b1bcc3dd02d6915bf9f6b745a7a20b9b112579e4c87fb5f4"} Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.775268 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ltn9g" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.778566 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zm6qq"] Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.779222 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vzljd" event={"ID":"c0706f0a-7f64-47d9-b3be-afe1d760603f","Type":"ContainerDied","Data":"c64f52522b823598007d99057217b63dc09189b0ca765150b4e30b4eb73180e1"} Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.779393 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vzljd" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.781710 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ltn9g" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.786708 4815 scope.go:117] "RemoveContainer" containerID="4384ecf6918b5b4aafdfe77e4f0a47f7836c43b8e0751493b2d47af69beb7b9b" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.797114 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zm6qq"] Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.820902 4815 scope.go:117] "RemoveContainer" containerID="a2379a4ffbdde16378a20d3074a26486558618bb4fca5b20693493f9d889e589" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.826951 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ltn9g" podStartSLOduration=1.826925321 podStartE2EDuration="1.826925321s" podCreationTimestamp="2026-02-25 13:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:25:54.821245186 +0000 UTC m=+312.622343250" watchObservedRunningTime="2026-02-25 13:25:54.826925321 +0000 UTC m=+312.628023385" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.841965 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6th6z"] Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.855214 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6th6z"] Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.864748 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6zz5"] Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.867895 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6zz5"] Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.872209 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vzljd"] Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.872648 4815 scope.go:117] "RemoveContainer" containerID="4eeb3c6973f20f0a56118b5dbf28db1a25436109bf0347e578acad431a1051b9" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.873478 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vzljd"] Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.876767 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9pjrm"] Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.881533 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9pjrm"] Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.902811 4815 scope.go:117] "RemoveContainer" containerID="a48ac1b49a391ebb803ae6d4669a0f80ae09a3be5f5ff2b349a1ba02e8f275fb" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.916001 4815 scope.go:117] "RemoveContainer" containerID="d8baf2b15896c3c7a59be66e98f35e71c42786196fb673aa0de2dc1894be3465" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.940805 4815 scope.go:117] "RemoveContainer" containerID="3ae1c93addc3743f6a3c4f03cb5cc5f88484ac039710085e2e9e90d700a80eac" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.944868 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b2fb1a6-e1d1-48e0-83dc-17143971294e" path="/var/lib/kubelet/pods/1b2fb1a6-e1d1-48e0-83dc-17143971294e/volumes" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.945707 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d00f8b4-1bf6-4a47-9beb-2226f26d695b" path="/var/lib/kubelet/pods/6d00f8b4-1bf6-4a47-9beb-2226f26d695b/volumes" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.946474 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c7db856-98cb-47a4-bad2-35ec76191138" path="/var/lib/kubelet/pods/9c7db856-98cb-47a4-bad2-35ec76191138/volumes" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.948196 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" path="/var/lib/kubelet/pods/bdf40b79-9414-4062-8dfd-3f0f71fcd1e4/volumes" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.948989 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0706f0a-7f64-47d9-b3be-afe1d760603f" path="/var/lib/kubelet/pods/c0706f0a-7f64-47d9-b3be-afe1d760603f/volumes" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.953620 4815 scope.go:117] "RemoveContainer" containerID="8248af535da5869c4f84a7e1453d2b519a4ff9c0ea2de89b618c2dcb5ea1040b" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.965193 4815 scope.go:117] "RemoveContainer" containerID="3f599ec4efa479f6413f95f683103e67ad256120d566559eded38b0a9cd092ec" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.979726 4815 scope.go:117] "RemoveContainer" containerID="5b2a897650c0e7a04106ae14a057313aef7644492d09282509472eee678d235c" Feb 25 13:25:54 crc kubenswrapper[4815]: I0225 13:25:54.991349 4815 scope.go:117] "RemoveContainer" containerID="ba6d7bc00ccdc194c823eea650bd84480e9545de9ba1158a9d9711d07c69862f" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.005088 4815 scope.go:117] "RemoveContainer" containerID="901fef9ce43607b5aee6845a1d8fe41620110fa225779efaa464a24d1bcd2f55" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.019308 4815 scope.go:117] "RemoveContainer" containerID="bd14764d84cfe21d3b0123f56abe3441bbb9f54618dcc498b0c3c007bfa74a03" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.638492 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b4jcd"] Feb 25 13:25:55 crc kubenswrapper[4815]: E0225 13:25:55.638691 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d00f8b4-1bf6-4a47-9beb-2226f26d695b" containerName="extract-content" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.638703 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d00f8b4-1bf6-4a47-9beb-2226f26d695b" containerName="extract-content" Feb 25 13:25:55 crc kubenswrapper[4815]: E0225 13:25:55.638712 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c7db856-98cb-47a4-bad2-35ec76191138" containerName="extract-utilities" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.638718 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c7db856-98cb-47a4-bad2-35ec76191138" containerName="extract-utilities" Feb 25 13:25:55 crc kubenswrapper[4815]: E0225 13:25:55.638727 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d00f8b4-1bf6-4a47-9beb-2226f26d695b" containerName="extract-utilities" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.638734 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d00f8b4-1bf6-4a47-9beb-2226f26d695b" containerName="extract-utilities" Feb 25 13:25:55 crc kubenswrapper[4815]: E0225 13:25:55.638743 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c7db856-98cb-47a4-bad2-35ec76191138" containerName="extract-content" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.638749 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c7db856-98cb-47a4-bad2-35ec76191138" containerName="extract-content" Feb 25 13:25:55 crc kubenswrapper[4815]: E0225 13:25:55.638760 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c7db856-98cb-47a4-bad2-35ec76191138" containerName="registry-server" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.638767 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c7db856-98cb-47a4-bad2-35ec76191138" containerName="registry-server" Feb 25 13:25:55 crc kubenswrapper[4815]: E0225 13:25:55.638775 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0706f0a-7f64-47d9-b3be-afe1d760603f" containerName="registry-server" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.638781 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0706f0a-7f64-47d9-b3be-afe1d760603f" containerName="registry-server" Feb 25 13:25:55 crc kubenswrapper[4815]: E0225 13:25:55.638792 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0706f0a-7f64-47d9-b3be-afe1d760603f" containerName="extract-utilities" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.638798 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0706f0a-7f64-47d9-b3be-afe1d760603f" containerName="extract-utilities" Feb 25 13:25:55 crc kubenswrapper[4815]: E0225 13:25:55.638804 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" containerName="registry-server" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.638810 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" containerName="registry-server" Feb 25 13:25:55 crc kubenswrapper[4815]: E0225 13:25:55.638818 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" containerName="extract-utilities" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.638824 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" containerName="extract-utilities" Feb 25 13:25:55 crc kubenswrapper[4815]: E0225 13:25:55.638833 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" containerName="extract-content" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.638838 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" containerName="extract-content" Feb 25 13:25:55 crc kubenswrapper[4815]: E0225 13:25:55.638844 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2fb1a6-e1d1-48e0-83dc-17143971294e" containerName="marketplace-operator" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.638850 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2fb1a6-e1d1-48e0-83dc-17143971294e" containerName="marketplace-operator" Feb 25 13:25:55 crc kubenswrapper[4815]: E0225 13:25:55.638857 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d00f8b4-1bf6-4a47-9beb-2226f26d695b" containerName="registry-server" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.638863 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d00f8b4-1bf6-4a47-9beb-2226f26d695b" containerName="registry-server" Feb 25 13:25:55 crc kubenswrapper[4815]: E0225 13:25:55.638874 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0706f0a-7f64-47d9-b3be-afe1d760603f" containerName="extract-content" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.638880 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0706f0a-7f64-47d9-b3be-afe1d760603f" containerName="extract-content" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.638980 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b2fb1a6-e1d1-48e0-83dc-17143971294e" containerName="marketplace-operator" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.638997 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c7db856-98cb-47a4-bad2-35ec76191138" containerName="registry-server" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.639004 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d00f8b4-1bf6-4a47-9beb-2226f26d695b" containerName="registry-server" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.639026 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0706f0a-7f64-47d9-b3be-afe1d760603f" containerName="registry-server" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.639033 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdf40b79-9414-4062-8dfd-3f0f71fcd1e4" containerName="registry-server" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.639710 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b4jcd" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.643070 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.647641 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b4jcd"] Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.715418 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdxsz\" (UniqueName: \"kubernetes.io/projected/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-kube-api-access-sdxsz\") pod \"certified-operators-b4jcd\" (UID: \"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc\") " pod="openshift-marketplace/certified-operators-b4jcd" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.715469 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-catalog-content\") pod \"certified-operators-b4jcd\" (UID: \"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc\") " pod="openshift-marketplace/certified-operators-b4jcd" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.715489 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-utilities\") pod \"certified-operators-b4jcd\" (UID: \"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc\") " pod="openshift-marketplace/certified-operators-b4jcd" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.817409 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdxsz\" (UniqueName: \"kubernetes.io/projected/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-kube-api-access-sdxsz\") pod \"certified-operators-b4jcd\" (UID: \"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc\") " pod="openshift-marketplace/certified-operators-b4jcd" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.817549 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-catalog-content\") pod \"certified-operators-b4jcd\" (UID: \"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc\") " pod="openshift-marketplace/certified-operators-b4jcd" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.817606 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-utilities\") pod \"certified-operators-b4jcd\" (UID: \"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc\") " pod="openshift-marketplace/certified-operators-b4jcd" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.820240 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-catalog-content\") pod \"certified-operators-b4jcd\" (UID: \"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc\") " pod="openshift-marketplace/certified-operators-b4jcd" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.821351 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-utilities\") pod \"certified-operators-b4jcd\" (UID: \"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc\") " pod="openshift-marketplace/certified-operators-b4jcd" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.831419 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b8shd"] Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.832414 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8shd" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.835672 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.842380 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8shd"] Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.843325 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdxsz\" (UniqueName: \"kubernetes.io/projected/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-kube-api-access-sdxsz\") pod \"certified-operators-b4jcd\" (UID: \"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc\") " pod="openshift-marketplace/certified-operators-b4jcd" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.919579 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06ec1b5e-fc51-4fe5-83d0-9697adaf1c57-catalog-content\") pod \"redhat-marketplace-b8shd\" (UID: \"06ec1b5e-fc51-4fe5-83d0-9697adaf1c57\") " pod="openshift-marketplace/redhat-marketplace-b8shd" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.919683 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06ec1b5e-fc51-4fe5-83d0-9697adaf1c57-utilities\") pod \"redhat-marketplace-b8shd\" (UID: \"06ec1b5e-fc51-4fe5-83d0-9697adaf1c57\") " pod="openshift-marketplace/redhat-marketplace-b8shd" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.919748 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5sr7\" (UniqueName: \"kubernetes.io/projected/06ec1b5e-fc51-4fe5-83d0-9697adaf1c57-kube-api-access-d5sr7\") pod \"redhat-marketplace-b8shd\" (UID: \"06ec1b5e-fc51-4fe5-83d0-9697adaf1c57\") " pod="openshift-marketplace/redhat-marketplace-b8shd" Feb 25 13:25:55 crc kubenswrapper[4815]: I0225 13:25:55.969694 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b4jcd" Feb 25 13:25:56 crc kubenswrapper[4815]: I0225 13:25:56.020985 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06ec1b5e-fc51-4fe5-83d0-9697adaf1c57-catalog-content\") pod \"redhat-marketplace-b8shd\" (UID: \"06ec1b5e-fc51-4fe5-83d0-9697adaf1c57\") " pod="openshift-marketplace/redhat-marketplace-b8shd" Feb 25 13:25:56 crc kubenswrapper[4815]: I0225 13:25:56.021072 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06ec1b5e-fc51-4fe5-83d0-9697adaf1c57-utilities\") pod \"redhat-marketplace-b8shd\" (UID: \"06ec1b5e-fc51-4fe5-83d0-9697adaf1c57\") " pod="openshift-marketplace/redhat-marketplace-b8shd" Feb 25 13:25:56 crc kubenswrapper[4815]: I0225 13:25:56.021103 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5sr7\" (UniqueName: \"kubernetes.io/projected/06ec1b5e-fc51-4fe5-83d0-9697adaf1c57-kube-api-access-d5sr7\") pod \"redhat-marketplace-b8shd\" (UID: \"06ec1b5e-fc51-4fe5-83d0-9697adaf1c57\") " pod="openshift-marketplace/redhat-marketplace-b8shd" Feb 25 13:25:56 crc kubenswrapper[4815]: I0225 13:25:56.021705 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06ec1b5e-fc51-4fe5-83d0-9697adaf1c57-catalog-content\") pod \"redhat-marketplace-b8shd\" (UID: \"06ec1b5e-fc51-4fe5-83d0-9697adaf1c57\") " pod="openshift-marketplace/redhat-marketplace-b8shd" Feb 25 13:25:56 crc kubenswrapper[4815]: I0225 13:25:56.021784 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06ec1b5e-fc51-4fe5-83d0-9697adaf1c57-utilities\") pod \"redhat-marketplace-b8shd\" (UID: \"06ec1b5e-fc51-4fe5-83d0-9697adaf1c57\") " pod="openshift-marketplace/redhat-marketplace-b8shd" Feb 25 13:25:56 crc kubenswrapper[4815]: I0225 13:25:56.040450 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5sr7\" (UniqueName: \"kubernetes.io/projected/06ec1b5e-fc51-4fe5-83d0-9697adaf1c57-kube-api-access-d5sr7\") pod \"redhat-marketplace-b8shd\" (UID: \"06ec1b5e-fc51-4fe5-83d0-9697adaf1c57\") " pod="openshift-marketplace/redhat-marketplace-b8shd" Feb 25 13:25:56 crc kubenswrapper[4815]: I0225 13:25:56.168087 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8shd" Feb 25 13:25:56 crc kubenswrapper[4815]: I0225 13:25:56.400925 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b4jcd"] Feb 25 13:25:56 crc kubenswrapper[4815]: W0225 13:25:56.403650 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6b25aff_3bb2_4fa7_9eea_ce5fc06b77cc.slice/crio-4af03c7c52770acd061f8e3f24c878ae5e4063e17d7b90fe55849fa540ab16a9 WatchSource:0}: Error finding container 4af03c7c52770acd061f8e3f24c878ae5e4063e17d7b90fe55849fa540ab16a9: Status 404 returned error can't find the container with id 4af03c7c52770acd061f8e3f24c878ae5e4063e17d7b90fe55849fa540ab16a9 Feb 25 13:25:56 crc kubenswrapper[4815]: I0225 13:25:56.592541 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8shd"] Feb 25 13:25:56 crc kubenswrapper[4815]: W0225 13:25:56.602098 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06ec1b5e_fc51_4fe5_83d0_9697adaf1c57.slice/crio-f145e7c6da0efe794bc77e582117a1da2ddbb23adf8ac0ad4a559f0dbbb79b11 WatchSource:0}: Error finding container f145e7c6da0efe794bc77e582117a1da2ddbb23adf8ac0ad4a559f0dbbb79b11: Status 404 returned error can't find the container with id f145e7c6da0efe794bc77e582117a1da2ddbb23adf8ac0ad4a559f0dbbb79b11 Feb 25 13:25:56 crc kubenswrapper[4815]: I0225 13:25:56.797501 4815 generic.go:334] "Generic (PLEG): container finished" podID="06ec1b5e-fc51-4fe5-83d0-9697adaf1c57" containerID="cf87c0c6fb7f38499f76c1ddb40a52815ffa766196f155ed1012f32d2c3a02f3" exitCode=0 Feb 25 13:25:56 crc kubenswrapper[4815]: I0225 13:25:56.797627 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8shd" event={"ID":"06ec1b5e-fc51-4fe5-83d0-9697adaf1c57","Type":"ContainerDied","Data":"cf87c0c6fb7f38499f76c1ddb40a52815ffa766196f155ed1012f32d2c3a02f3"} Feb 25 13:25:56 crc kubenswrapper[4815]: I0225 13:25:56.797823 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8shd" event={"ID":"06ec1b5e-fc51-4fe5-83d0-9697adaf1c57","Type":"ContainerStarted","Data":"f145e7c6da0efe794bc77e582117a1da2ddbb23adf8ac0ad4a559f0dbbb79b11"} Feb 25 13:25:56 crc kubenswrapper[4815]: I0225 13:25:56.802271 4815 generic.go:334] "Generic (PLEG): container finished" podID="f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc" containerID="a4f0c05b7bfba0f138faed2f793181e370c1507683619e0a61e96199c88149f7" exitCode=0 Feb 25 13:25:56 crc kubenswrapper[4815]: I0225 13:25:56.803135 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4jcd" event={"ID":"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc","Type":"ContainerDied","Data":"a4f0c05b7bfba0f138faed2f793181e370c1507683619e0a61e96199c88149f7"} Feb 25 13:25:56 crc kubenswrapper[4815]: I0225 13:25:56.803177 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4jcd" event={"ID":"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc","Type":"ContainerStarted","Data":"4af03c7c52770acd061f8e3f24c878ae5e4063e17d7b90fe55849fa540ab16a9"} Feb 25 13:25:57 crc kubenswrapper[4815]: I0225 13:25:57.815299 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4jcd" event={"ID":"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc","Type":"ContainerStarted","Data":"9a5e6da638c04a2b0223119a8c05aafe33d9dcff08004fd86b264dc6add5a6fc"} Feb 25 13:25:57 crc kubenswrapper[4815]: I0225 13:25:57.818908 4815 generic.go:334] "Generic (PLEG): container finished" podID="06ec1b5e-fc51-4fe5-83d0-9697adaf1c57" containerID="7fdea427ec1bc4d9a88a44ca3d88da1584f7e5b1332f42f7fde018c3df65538f" exitCode=0 Feb 25 13:25:57 crc kubenswrapper[4815]: I0225 13:25:57.818963 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8shd" event={"ID":"06ec1b5e-fc51-4fe5-83d0-9697adaf1c57","Type":"ContainerDied","Data":"7fdea427ec1bc4d9a88a44ca3d88da1584f7e5b1332f42f7fde018c3df65538f"} Feb 25 13:25:57 crc kubenswrapper[4815]: E0225 13:25:57.913476 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6b25aff_3bb2_4fa7_9eea_ce5fc06b77cc.slice/crio-9a5e6da638c04a2b0223119a8c05aafe33d9dcff08004fd86b264dc6add5a6fc.scope\": RecentStats: unable to find data in memory cache]" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.045038 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xw2zf"] Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.046395 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xw2zf" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.048866 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.050318 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xw2zf"] Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.150539 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9db66c63-07f2-4e45-851e-108477a7159d-catalog-content\") pod \"redhat-operators-xw2zf\" (UID: \"9db66c63-07f2-4e45-851e-108477a7159d\") " pod="openshift-marketplace/redhat-operators-xw2zf" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.150713 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrmgk\" (UniqueName: \"kubernetes.io/projected/9db66c63-07f2-4e45-851e-108477a7159d-kube-api-access-zrmgk\") pod \"redhat-operators-xw2zf\" (UID: \"9db66c63-07f2-4e45-851e-108477a7159d\") " pod="openshift-marketplace/redhat-operators-xw2zf" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.150838 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9db66c63-07f2-4e45-851e-108477a7159d-utilities\") pod \"redhat-operators-xw2zf\" (UID: \"9db66c63-07f2-4e45-851e-108477a7159d\") " pod="openshift-marketplace/redhat-operators-xw2zf" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.229026 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mvxkh"] Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.229955 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvxkh" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.232334 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.239055 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mvxkh"] Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.252062 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrmgk\" (UniqueName: \"kubernetes.io/projected/9db66c63-07f2-4e45-851e-108477a7159d-kube-api-access-zrmgk\") pod \"redhat-operators-xw2zf\" (UID: \"9db66c63-07f2-4e45-851e-108477a7159d\") " pod="openshift-marketplace/redhat-operators-xw2zf" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.252123 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9db66c63-07f2-4e45-851e-108477a7159d-utilities\") pod \"redhat-operators-xw2zf\" (UID: \"9db66c63-07f2-4e45-851e-108477a7159d\") " pod="openshift-marketplace/redhat-operators-xw2zf" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.252156 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9db66c63-07f2-4e45-851e-108477a7159d-catalog-content\") pod \"redhat-operators-xw2zf\" (UID: \"9db66c63-07f2-4e45-851e-108477a7159d\") " pod="openshift-marketplace/redhat-operators-xw2zf" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.252605 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9db66c63-07f2-4e45-851e-108477a7159d-catalog-content\") pod \"redhat-operators-xw2zf\" (UID: \"9db66c63-07f2-4e45-851e-108477a7159d\") " pod="openshift-marketplace/redhat-operators-xw2zf" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.253042 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9db66c63-07f2-4e45-851e-108477a7159d-utilities\") pod \"redhat-operators-xw2zf\" (UID: \"9db66c63-07f2-4e45-851e-108477a7159d\") " pod="openshift-marketplace/redhat-operators-xw2zf" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.270731 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrmgk\" (UniqueName: \"kubernetes.io/projected/9db66c63-07f2-4e45-851e-108477a7159d-kube-api-access-zrmgk\") pod \"redhat-operators-xw2zf\" (UID: \"9db66c63-07f2-4e45-851e-108477a7159d\") " pod="openshift-marketplace/redhat-operators-xw2zf" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.352810 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fae8512-7a85-4d34-87f3-d4d2e6832d93-catalog-content\") pod \"community-operators-mvxkh\" (UID: \"7fae8512-7a85-4d34-87f3-d4d2e6832d93\") " pod="openshift-marketplace/community-operators-mvxkh" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.352896 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldjqw\" (UniqueName: \"kubernetes.io/projected/7fae8512-7a85-4d34-87f3-d4d2e6832d93-kube-api-access-ldjqw\") pod \"community-operators-mvxkh\" (UID: \"7fae8512-7a85-4d34-87f3-d4d2e6832d93\") " pod="openshift-marketplace/community-operators-mvxkh" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.352947 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fae8512-7a85-4d34-87f3-d4d2e6832d93-utilities\") pod \"community-operators-mvxkh\" (UID: \"7fae8512-7a85-4d34-87f3-d4d2e6832d93\") " pod="openshift-marketplace/community-operators-mvxkh" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.368633 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xw2zf" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.455075 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fae8512-7a85-4d34-87f3-d4d2e6832d93-utilities\") pod \"community-operators-mvxkh\" (UID: \"7fae8512-7a85-4d34-87f3-d4d2e6832d93\") " pod="openshift-marketplace/community-operators-mvxkh" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.455419 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fae8512-7a85-4d34-87f3-d4d2e6832d93-catalog-content\") pod \"community-operators-mvxkh\" (UID: \"7fae8512-7a85-4d34-87f3-d4d2e6832d93\") " pod="openshift-marketplace/community-operators-mvxkh" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.455484 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldjqw\" (UniqueName: \"kubernetes.io/projected/7fae8512-7a85-4d34-87f3-d4d2e6832d93-kube-api-access-ldjqw\") pod \"community-operators-mvxkh\" (UID: \"7fae8512-7a85-4d34-87f3-d4d2e6832d93\") " pod="openshift-marketplace/community-operators-mvxkh" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.455956 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fae8512-7a85-4d34-87f3-d4d2e6832d93-utilities\") pod \"community-operators-mvxkh\" (UID: \"7fae8512-7a85-4d34-87f3-d4d2e6832d93\") " pod="openshift-marketplace/community-operators-mvxkh" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.456023 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fae8512-7a85-4d34-87f3-d4d2e6832d93-catalog-content\") pod \"community-operators-mvxkh\" (UID: \"7fae8512-7a85-4d34-87f3-d4d2e6832d93\") " pod="openshift-marketplace/community-operators-mvxkh" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.476165 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldjqw\" (UniqueName: \"kubernetes.io/projected/7fae8512-7a85-4d34-87f3-d4d2e6832d93-kube-api-access-ldjqw\") pod \"community-operators-mvxkh\" (UID: \"7fae8512-7a85-4d34-87f3-d4d2e6832d93\") " pod="openshift-marketplace/community-operators-mvxkh" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.598538 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvxkh" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.760088 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xw2zf"] Feb 25 13:25:58 crc kubenswrapper[4815]: W0225 13:25:58.769372 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9db66c63_07f2_4e45_851e_108477a7159d.slice/crio-b57e82c9b968302a0f91ede85473fcf02692a82a130ba51f95c15afc8f119aaa WatchSource:0}: Error finding container b57e82c9b968302a0f91ede85473fcf02692a82a130ba51f95c15afc8f119aaa: Status 404 returned error can't find the container with id b57e82c9b968302a0f91ede85473fcf02692a82a130ba51f95c15afc8f119aaa Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.827216 4815 generic.go:334] "Generic (PLEG): container finished" podID="f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc" containerID="9a5e6da638c04a2b0223119a8c05aafe33d9dcff08004fd86b264dc6add5a6fc" exitCode=0 Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.827559 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4jcd" event={"ID":"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc","Type":"ContainerDied","Data":"9a5e6da638c04a2b0223119a8c05aafe33d9dcff08004fd86b264dc6add5a6fc"} Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.830818 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xw2zf" event={"ID":"9db66c63-07f2-4e45-851e-108477a7159d","Type":"ContainerStarted","Data":"b57e82c9b968302a0f91ede85473fcf02692a82a130ba51f95c15afc8f119aaa"} Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.839814 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8shd" event={"ID":"06ec1b5e-fc51-4fe5-83d0-9697adaf1c57","Type":"ContainerStarted","Data":"c1e482ce310d1c7d9e7947eaf7f3ad1491a42feba762979875c8e55f96ea80c9"} Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.871931 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b8shd" podStartSLOduration=2.461181205 podStartE2EDuration="3.871909495s" podCreationTimestamp="2026-02-25 13:25:55 +0000 UTC" firstStartedPulling="2026-02-25 13:25:56.798840526 +0000 UTC m=+314.599938580" lastFinishedPulling="2026-02-25 13:25:58.209568816 +0000 UTC m=+316.010666870" observedRunningTime="2026-02-25 13:25:58.86561893 +0000 UTC m=+316.666716994" watchObservedRunningTime="2026-02-25 13:25:58.871909495 +0000 UTC m=+316.673007549" Feb 25 13:25:58 crc kubenswrapper[4815]: I0225 13:25:58.973237 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mvxkh"] Feb 25 13:25:58 crc kubenswrapper[4815]: W0225 13:25:58.981219 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fae8512_7a85_4d34_87f3_d4d2e6832d93.slice/crio-eb15f7573a805fe3c98ed29843d8799a0c5ffbdec3536975882964e7b51ab0b1 WatchSource:0}: Error finding container eb15f7573a805fe3c98ed29843d8799a0c5ffbdec3536975882964e7b51ab0b1: Status 404 returned error can't find the container with id eb15f7573a805fe3c98ed29843d8799a0c5ffbdec3536975882964e7b51ab0b1 Feb 25 13:25:59 crc kubenswrapper[4815]: I0225 13:25:59.847194 4815 generic.go:334] "Generic (PLEG): container finished" podID="7fae8512-7a85-4d34-87f3-d4d2e6832d93" containerID="06afe600018f51c39a204ead37fdd6a4e0440a72377785d57ecaf1458e955962" exitCode=0 Feb 25 13:25:59 crc kubenswrapper[4815]: I0225 13:25:59.847485 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvxkh" event={"ID":"7fae8512-7a85-4d34-87f3-d4d2e6832d93","Type":"ContainerDied","Data":"06afe600018f51c39a204ead37fdd6a4e0440a72377785d57ecaf1458e955962"} Feb 25 13:25:59 crc kubenswrapper[4815]: I0225 13:25:59.847620 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvxkh" event={"ID":"7fae8512-7a85-4d34-87f3-d4d2e6832d93","Type":"ContainerStarted","Data":"eb15f7573a805fe3c98ed29843d8799a0c5ffbdec3536975882964e7b51ab0b1"} Feb 25 13:25:59 crc kubenswrapper[4815]: I0225 13:25:59.851397 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4jcd" event={"ID":"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc","Type":"ContainerStarted","Data":"f94fd26d90a1ba7ce5481ebdce566b03cc4fa187919a2817bdad927c880a5288"} Feb 25 13:25:59 crc kubenswrapper[4815]: I0225 13:25:59.853991 4815 generic.go:334] "Generic (PLEG): container finished" podID="9db66c63-07f2-4e45-851e-108477a7159d" containerID="d3199b3cb67cfa727d84006c3f30e4b17c2dcf0e6817df4cc81ea9df173bf4fa" exitCode=0 Feb 25 13:25:59 crc kubenswrapper[4815]: I0225 13:25:59.854060 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xw2zf" event={"ID":"9db66c63-07f2-4e45-851e-108477a7159d","Type":"ContainerDied","Data":"d3199b3cb67cfa727d84006c3f30e4b17c2dcf0e6817df4cc81ea9df173bf4fa"} Feb 25 13:25:59 crc kubenswrapper[4815]: I0225 13:25:59.906702 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b4jcd" podStartSLOduration=2.476976532 podStartE2EDuration="4.906676582s" podCreationTimestamp="2026-02-25 13:25:55 +0000 UTC" firstStartedPulling="2026-02-25 13:25:56.804972457 +0000 UTC m=+314.606070511" lastFinishedPulling="2026-02-25 13:25:59.234672507 +0000 UTC m=+317.035770561" observedRunningTime="2026-02-25 13:25:59.899543899 +0000 UTC m=+317.700641973" watchObservedRunningTime="2026-02-25 13:25:59.906676582 +0000 UTC m=+317.707774656" Feb 25 13:26:00 crc kubenswrapper[4815]: I0225 13:26:00.129814 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533766-d62n6"] Feb 25 13:26:00 crc kubenswrapper[4815]: I0225 13:26:00.130725 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533766-d62n6" Feb 25 13:26:00 crc kubenswrapper[4815]: I0225 13:26:00.135200 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533766-d62n6"] Feb 25 13:26:00 crc kubenswrapper[4815]: I0225 13:26:00.136330 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:26:00 crc kubenswrapper[4815]: I0225 13:26:00.136453 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:26:00 crc kubenswrapper[4815]: I0225 13:26:00.137470 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:26:00 crc kubenswrapper[4815]: I0225 13:26:00.215356 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khksp\" (UniqueName: \"kubernetes.io/projected/5bfd8350-d886-4279-aac6-28e1f1c1f752-kube-api-access-khksp\") pod \"auto-csr-approver-29533766-d62n6\" (UID: \"5bfd8350-d886-4279-aac6-28e1f1c1f752\") " pod="openshift-infra/auto-csr-approver-29533766-d62n6" Feb 25 13:26:00 crc kubenswrapper[4815]: I0225 13:26:00.316702 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khksp\" (UniqueName: \"kubernetes.io/projected/5bfd8350-d886-4279-aac6-28e1f1c1f752-kube-api-access-khksp\") pod \"auto-csr-approver-29533766-d62n6\" (UID: \"5bfd8350-d886-4279-aac6-28e1f1c1f752\") " pod="openshift-infra/auto-csr-approver-29533766-d62n6" Feb 25 13:26:00 crc kubenswrapper[4815]: I0225 13:26:00.342844 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khksp\" (UniqueName: \"kubernetes.io/projected/5bfd8350-d886-4279-aac6-28e1f1c1f752-kube-api-access-khksp\") pod \"auto-csr-approver-29533766-d62n6\" (UID: \"5bfd8350-d886-4279-aac6-28e1f1c1f752\") " pod="openshift-infra/auto-csr-approver-29533766-d62n6" Feb 25 13:26:00 crc kubenswrapper[4815]: I0225 13:26:00.445918 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533766-d62n6" Feb 25 13:26:00 crc kubenswrapper[4815]: I0225 13:26:00.672583 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533766-d62n6"] Feb 25 13:26:00 crc kubenswrapper[4815]: I0225 13:26:00.862105 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvxkh" event={"ID":"7fae8512-7a85-4d34-87f3-d4d2e6832d93","Type":"ContainerStarted","Data":"2895998dbd5e494f739e100bda13cd2b777b28eec6f42ee38dc5f44c4baadf38"} Feb 25 13:26:00 crc kubenswrapper[4815]: I0225 13:26:00.870383 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xw2zf" event={"ID":"9db66c63-07f2-4e45-851e-108477a7159d","Type":"ContainerStarted","Data":"627848e2c735399831f64c26ac6b00e6aec7276e7d60f3fd0bce07b22f595b14"} Feb 25 13:26:00 crc kubenswrapper[4815]: I0225 13:26:00.871493 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533766-d62n6" event={"ID":"5bfd8350-d886-4279-aac6-28e1f1c1f752","Type":"ContainerStarted","Data":"98a15df10aeb9d5c051c1cf032e602473f5e0b0ffe106f4ace039abe69f09f5b"} Feb 25 13:26:01 crc kubenswrapper[4815]: I0225 13:26:01.895420 4815 generic.go:334] "Generic (PLEG): container finished" podID="9db66c63-07f2-4e45-851e-108477a7159d" containerID="627848e2c735399831f64c26ac6b00e6aec7276e7d60f3fd0bce07b22f595b14" exitCode=0 Feb 25 13:26:01 crc kubenswrapper[4815]: I0225 13:26:01.895480 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xw2zf" event={"ID":"9db66c63-07f2-4e45-851e-108477a7159d","Type":"ContainerDied","Data":"627848e2c735399831f64c26ac6b00e6aec7276e7d60f3fd0bce07b22f595b14"} Feb 25 13:26:01 crc kubenswrapper[4815]: I0225 13:26:01.898959 4815 generic.go:334] "Generic (PLEG): container finished" podID="7fae8512-7a85-4d34-87f3-d4d2e6832d93" containerID="2895998dbd5e494f739e100bda13cd2b777b28eec6f42ee38dc5f44c4baadf38" exitCode=0 Feb 25 13:26:01 crc kubenswrapper[4815]: I0225 13:26:01.898998 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvxkh" event={"ID":"7fae8512-7a85-4d34-87f3-d4d2e6832d93","Type":"ContainerDied","Data":"2895998dbd5e494f739e100bda13cd2b777b28eec6f42ee38dc5f44c4baadf38"} Feb 25 13:26:02 crc kubenswrapper[4815]: I0225 13:26:02.906914 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xw2zf" event={"ID":"9db66c63-07f2-4e45-851e-108477a7159d","Type":"ContainerStarted","Data":"dff6e77aeaefc637e1d4e62f11254e40e11413522a67ffbc1730bc4eae435f95"} Feb 25 13:26:02 crc kubenswrapper[4815]: I0225 13:26:02.908110 4815 generic.go:334] "Generic (PLEG): container finished" podID="5bfd8350-d886-4279-aac6-28e1f1c1f752" containerID="e8da71b1f37781e767db03771dca29938182875e21753e4c582698e896a74a8a" exitCode=0 Feb 25 13:26:02 crc kubenswrapper[4815]: I0225 13:26:02.908159 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533766-d62n6" event={"ID":"5bfd8350-d886-4279-aac6-28e1f1c1f752","Type":"ContainerDied","Data":"e8da71b1f37781e767db03771dca29938182875e21753e4c582698e896a74a8a"} Feb 25 13:26:02 crc kubenswrapper[4815]: I0225 13:26:02.909929 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvxkh" event={"ID":"7fae8512-7a85-4d34-87f3-d4d2e6832d93","Type":"ContainerStarted","Data":"addc89334739906dfdeed0b3362af2d90ea3537c14f3d8cf7cd4a20c2d24ea52"} Feb 25 13:26:02 crc kubenswrapper[4815]: I0225 13:26:02.965081 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xw2zf" podStartSLOduration=2.327945935 podStartE2EDuration="4.965061432s" podCreationTimestamp="2026-02-25 13:25:58 +0000 UTC" firstStartedPulling="2026-02-25 13:25:59.855294363 +0000 UTC m=+317.656392437" lastFinishedPulling="2026-02-25 13:26:02.49240988 +0000 UTC m=+320.293507934" observedRunningTime="2026-02-25 13:26:02.932816189 +0000 UTC m=+320.733914243" watchObservedRunningTime="2026-02-25 13:26:02.965061432 +0000 UTC m=+320.766159486" Feb 25 13:26:02 crc kubenswrapper[4815]: I0225 13:26:02.984242 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mvxkh" podStartSLOduration=2.489114072 podStartE2EDuration="4.984223589s" podCreationTimestamp="2026-02-25 13:25:58 +0000 UTC" firstStartedPulling="2026-02-25 13:25:59.848822802 +0000 UTC m=+317.649920866" lastFinishedPulling="2026-02-25 13:26:02.343932329 +0000 UTC m=+320.145030383" observedRunningTime="2026-02-25 13:26:02.981648054 +0000 UTC m=+320.782746118" watchObservedRunningTime="2026-02-25 13:26:02.984223589 +0000 UTC m=+320.785321633" Feb 25 13:26:04 crc kubenswrapper[4815]: I0225 13:26:04.194480 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533766-d62n6" Feb 25 13:26:04 crc kubenswrapper[4815]: I0225 13:26:04.374810 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khksp\" (UniqueName: \"kubernetes.io/projected/5bfd8350-d886-4279-aac6-28e1f1c1f752-kube-api-access-khksp\") pod \"5bfd8350-d886-4279-aac6-28e1f1c1f752\" (UID: \"5bfd8350-d886-4279-aac6-28e1f1c1f752\") " Feb 25 13:26:04 crc kubenswrapper[4815]: I0225 13:26:04.380158 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bfd8350-d886-4279-aac6-28e1f1c1f752-kube-api-access-khksp" (OuterVolumeSpecName: "kube-api-access-khksp") pod "5bfd8350-d886-4279-aac6-28e1f1c1f752" (UID: "5bfd8350-d886-4279-aac6-28e1f1c1f752"). InnerVolumeSpecName "kube-api-access-khksp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:26:04 crc kubenswrapper[4815]: I0225 13:26:04.476339 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khksp\" (UniqueName: \"kubernetes.io/projected/5bfd8350-d886-4279-aac6-28e1f1c1f752-kube-api-access-khksp\") on node \"crc\" DevicePath \"\"" Feb 25 13:26:04 crc kubenswrapper[4815]: I0225 13:26:04.920614 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533766-d62n6" event={"ID":"5bfd8350-d886-4279-aac6-28e1f1c1f752","Type":"ContainerDied","Data":"98a15df10aeb9d5c051c1cf032e602473f5e0b0ffe106f4ace039abe69f09f5b"} Feb 25 13:26:04 crc kubenswrapper[4815]: I0225 13:26:04.920652 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98a15df10aeb9d5c051c1cf032e602473f5e0b0ffe106f4ace039abe69f09f5b" Feb 25 13:26:04 crc kubenswrapper[4815]: I0225 13:26:04.920668 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533766-d62n6" Feb 25 13:26:05 crc kubenswrapper[4815]: I0225 13:26:05.971308 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b4jcd" Feb 25 13:26:05 crc kubenswrapper[4815]: I0225 13:26:05.971666 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b4jcd" Feb 25 13:26:06 crc kubenswrapper[4815]: I0225 13:26:06.025052 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b4jcd" Feb 25 13:26:06 crc kubenswrapper[4815]: I0225 13:26:06.169093 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b8shd" Feb 25 13:26:06 crc kubenswrapper[4815]: I0225 13:26:06.169184 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b8shd" Feb 25 13:26:06 crc kubenswrapper[4815]: I0225 13:26:06.211006 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b8shd" Feb 25 13:26:06 crc kubenswrapper[4815]: I0225 13:26:06.974423 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b4jcd" Feb 25 13:26:06 crc kubenswrapper[4815]: I0225 13:26:06.977877 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b8shd" Feb 25 13:26:08 crc kubenswrapper[4815]: I0225 13:26:08.369341 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xw2zf" Feb 25 13:26:08 crc kubenswrapper[4815]: I0225 13:26:08.369733 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xw2zf" Feb 25 13:26:08 crc kubenswrapper[4815]: I0225 13:26:08.422943 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xw2zf" Feb 25 13:26:08 crc kubenswrapper[4815]: I0225 13:26:08.598862 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mvxkh" Feb 25 13:26:08 crc kubenswrapper[4815]: I0225 13:26:08.598929 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mvxkh" Feb 25 13:26:08 crc kubenswrapper[4815]: I0225 13:26:08.643233 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mvxkh" Feb 25 13:26:09 crc kubenswrapper[4815]: I0225 13:26:09.016080 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mvxkh" Feb 25 13:26:09 crc kubenswrapper[4815]: I0225 13:26:09.021014 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xw2zf" Feb 25 13:27:21 crc kubenswrapper[4815]: I0225 13:27:21.342583 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:27:21 crc kubenswrapper[4815]: I0225 13:27:21.343703 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:27:51 crc kubenswrapper[4815]: I0225 13:27:51.342365 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:27:51 crc kubenswrapper[4815]: I0225 13:27:51.343138 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:28:00 crc kubenswrapper[4815]: I0225 13:28:00.137484 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533768-b4gbs"] Feb 25 13:28:00 crc kubenswrapper[4815]: E0225 13:28:00.138233 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bfd8350-d886-4279-aac6-28e1f1c1f752" containerName="oc" Feb 25 13:28:00 crc kubenswrapper[4815]: I0225 13:28:00.138248 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bfd8350-d886-4279-aac6-28e1f1c1f752" containerName="oc" Feb 25 13:28:00 crc kubenswrapper[4815]: I0225 13:28:00.138416 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bfd8350-d886-4279-aac6-28e1f1c1f752" containerName="oc" Feb 25 13:28:00 crc kubenswrapper[4815]: I0225 13:28:00.138837 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533768-b4gbs" Feb 25 13:28:00 crc kubenswrapper[4815]: I0225 13:28:00.141979 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:28:00 crc kubenswrapper[4815]: I0225 13:28:00.142114 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:28:00 crc kubenswrapper[4815]: I0225 13:28:00.142334 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:28:00 crc kubenswrapper[4815]: I0225 13:28:00.153620 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533768-b4gbs"] Feb 25 13:28:00 crc kubenswrapper[4815]: I0225 13:28:00.172224 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znlxs\" (UniqueName: \"kubernetes.io/projected/02dfab6b-ba3d-4288-b8f2-e41422e7cc78-kube-api-access-znlxs\") pod \"auto-csr-approver-29533768-b4gbs\" (UID: \"02dfab6b-ba3d-4288-b8f2-e41422e7cc78\") " pod="openshift-infra/auto-csr-approver-29533768-b4gbs" Feb 25 13:28:00 crc kubenswrapper[4815]: I0225 13:28:00.274282 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znlxs\" (UniqueName: \"kubernetes.io/projected/02dfab6b-ba3d-4288-b8f2-e41422e7cc78-kube-api-access-znlxs\") pod \"auto-csr-approver-29533768-b4gbs\" (UID: \"02dfab6b-ba3d-4288-b8f2-e41422e7cc78\") " pod="openshift-infra/auto-csr-approver-29533768-b4gbs" Feb 25 13:28:00 crc kubenswrapper[4815]: I0225 13:28:00.293257 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znlxs\" (UniqueName: \"kubernetes.io/projected/02dfab6b-ba3d-4288-b8f2-e41422e7cc78-kube-api-access-znlxs\") pod \"auto-csr-approver-29533768-b4gbs\" (UID: \"02dfab6b-ba3d-4288-b8f2-e41422e7cc78\") " pod="openshift-infra/auto-csr-approver-29533768-b4gbs" Feb 25 13:28:00 crc kubenswrapper[4815]: I0225 13:28:00.459644 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533768-b4gbs" Feb 25 13:28:00 crc kubenswrapper[4815]: I0225 13:28:00.686857 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533768-b4gbs"] Feb 25 13:28:00 crc kubenswrapper[4815]: I0225 13:28:00.696893 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 13:28:00 crc kubenswrapper[4815]: I0225 13:28:00.702837 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533768-b4gbs" event={"ID":"02dfab6b-ba3d-4288-b8f2-e41422e7cc78","Type":"ContainerStarted","Data":"4b53d952d6a665198aea18f8bf02737a020fde44f8649fe1d57e4f7f1dca3475"} Feb 25 13:28:02 crc kubenswrapper[4815]: I0225 13:28:02.720197 4815 generic.go:334] "Generic (PLEG): container finished" podID="02dfab6b-ba3d-4288-b8f2-e41422e7cc78" containerID="d436db49705ce47a5ea2e0f27945c303583256e6b301f477c0ce993911b28b73" exitCode=0 Feb 25 13:28:02 crc kubenswrapper[4815]: I0225 13:28:02.720301 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533768-b4gbs" event={"ID":"02dfab6b-ba3d-4288-b8f2-e41422e7cc78","Type":"ContainerDied","Data":"d436db49705ce47a5ea2e0f27945c303583256e6b301f477c0ce993911b28b73"} Feb 25 13:28:03 crc kubenswrapper[4815]: I0225 13:28:03.955979 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533768-b4gbs" Feb 25 13:28:04 crc kubenswrapper[4815]: I0225 13:28:04.119372 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znlxs\" (UniqueName: \"kubernetes.io/projected/02dfab6b-ba3d-4288-b8f2-e41422e7cc78-kube-api-access-znlxs\") pod \"02dfab6b-ba3d-4288-b8f2-e41422e7cc78\" (UID: \"02dfab6b-ba3d-4288-b8f2-e41422e7cc78\") " Feb 25 13:28:04 crc kubenswrapper[4815]: I0225 13:28:04.126279 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02dfab6b-ba3d-4288-b8f2-e41422e7cc78-kube-api-access-znlxs" (OuterVolumeSpecName: "kube-api-access-znlxs") pod "02dfab6b-ba3d-4288-b8f2-e41422e7cc78" (UID: "02dfab6b-ba3d-4288-b8f2-e41422e7cc78"). InnerVolumeSpecName "kube-api-access-znlxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:28:04 crc kubenswrapper[4815]: I0225 13:28:04.220889 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znlxs\" (UniqueName: \"kubernetes.io/projected/02dfab6b-ba3d-4288-b8f2-e41422e7cc78-kube-api-access-znlxs\") on node \"crc\" DevicePath \"\"" Feb 25 13:28:04 crc kubenswrapper[4815]: I0225 13:28:04.734623 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533768-b4gbs" event={"ID":"02dfab6b-ba3d-4288-b8f2-e41422e7cc78","Type":"ContainerDied","Data":"4b53d952d6a665198aea18f8bf02737a020fde44f8649fe1d57e4f7f1dca3475"} Feb 25 13:28:04 crc kubenswrapper[4815]: I0225 13:28:04.734681 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b53d952d6a665198aea18f8bf02737a020fde44f8649fe1d57e4f7f1dca3475" Feb 25 13:28:04 crc kubenswrapper[4815]: I0225 13:28:04.734734 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533768-b4gbs" Feb 25 13:28:05 crc kubenswrapper[4815]: I0225 13:28:05.016127 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533762-lkzv9"] Feb 25 13:28:05 crc kubenswrapper[4815]: I0225 13:28:05.020284 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533762-lkzv9"] Feb 25 13:28:06 crc kubenswrapper[4815]: I0225 13:28:06.943746 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6403af09-14ef-49fc-858f-06c1bbadb88b" path="/var/lib/kubelet/pods/6403af09-14ef-49fc-858f-06c1bbadb88b/volumes" Feb 25 13:28:21 crc kubenswrapper[4815]: I0225 13:28:21.342348 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:28:21 crc kubenswrapper[4815]: I0225 13:28:21.342879 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:28:21 crc kubenswrapper[4815]: I0225 13:28:21.342978 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:28:21 crc kubenswrapper[4815]: I0225 13:28:21.344052 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0bde58aaffd292f31f5aeec24700d4368aa27e806f31f168fd069eb408b17058"} pod="openshift-machine-config-operator/machine-config-daemon-frc75" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 13:28:21 crc kubenswrapper[4815]: I0225 13:28:21.344172 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" containerID="cri-o://0bde58aaffd292f31f5aeec24700d4368aa27e806f31f168fd069eb408b17058" gracePeriod=600 Feb 25 13:28:21 crc kubenswrapper[4815]: I0225 13:28:21.841133 4815 generic.go:334] "Generic (PLEG): container finished" podID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerID="0bde58aaffd292f31f5aeec24700d4368aa27e806f31f168fd069eb408b17058" exitCode=0 Feb 25 13:28:21 crc kubenswrapper[4815]: I0225 13:28:21.841307 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerDied","Data":"0bde58aaffd292f31f5aeec24700d4368aa27e806f31f168fd069eb408b17058"} Feb 25 13:28:21 crc kubenswrapper[4815]: I0225 13:28:21.841842 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerStarted","Data":"4e00e443cf7572407e0857913be27d833d029fb2c70b14c349ec691336c7312d"} Feb 25 13:28:21 crc kubenswrapper[4815]: I0225 13:28:21.841915 4815 scope.go:117] "RemoveContainer" containerID="440b745816f3d3542a46a14793d9d030a841eaa34e91cd17885d1961ecc85bee" Feb 25 13:29:43 crc kubenswrapper[4815]: I0225 13:29:43.305568 4815 scope.go:117] "RemoveContainer" containerID="92c3363b249ba9f79180012961834296c8953a14c98c4d67183a8e953ed394f6" Feb 25 13:29:43 crc kubenswrapper[4815]: I0225 13:29:43.336140 4815 scope.go:117] "RemoveContainer" containerID="fc9e0681c3d6bd9c78d893b8852443115a507df82241b05734da777a03a3a18d" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.151830 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533770-qrbhh"] Feb 25 13:30:00 crc kubenswrapper[4815]: E0225 13:30:00.152981 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02dfab6b-ba3d-4288-b8f2-e41422e7cc78" containerName="oc" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.153009 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="02dfab6b-ba3d-4288-b8f2-e41422e7cc78" containerName="oc" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.153263 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="02dfab6b-ba3d-4288-b8f2-e41422e7cc78" containerName="oc" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.154051 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533770-qrbhh" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.157074 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt"] Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.157401 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.157980 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.158457 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.160723 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.163663 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.163697 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.171249 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533770-qrbhh"] Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.175810 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt"] Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.277663 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6f07757-816a-4dd2-b804-795dd6a874fd-secret-volume\") pod \"collect-profiles-29533770-sjpgt\" (UID: \"b6f07757-816a-4dd2-b804-795dd6a874fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.278104 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6f07757-816a-4dd2-b804-795dd6a874fd-config-volume\") pod \"collect-profiles-29533770-sjpgt\" (UID: \"b6f07757-816a-4dd2-b804-795dd6a874fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.278314 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsrkz\" (UniqueName: \"kubernetes.io/projected/67c273aa-5ff2-43fc-871f-64caafce5463-kube-api-access-xsrkz\") pod \"auto-csr-approver-29533770-qrbhh\" (UID: \"67c273aa-5ff2-43fc-871f-64caafce5463\") " pod="openshift-infra/auto-csr-approver-29533770-qrbhh" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.278470 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7knm\" (UniqueName: \"kubernetes.io/projected/b6f07757-816a-4dd2-b804-795dd6a874fd-kube-api-access-v7knm\") pod \"collect-profiles-29533770-sjpgt\" (UID: \"b6f07757-816a-4dd2-b804-795dd6a874fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.380190 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsrkz\" (UniqueName: \"kubernetes.io/projected/67c273aa-5ff2-43fc-871f-64caafce5463-kube-api-access-xsrkz\") pod \"auto-csr-approver-29533770-qrbhh\" (UID: \"67c273aa-5ff2-43fc-871f-64caafce5463\") " pod="openshift-infra/auto-csr-approver-29533770-qrbhh" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.380271 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7knm\" (UniqueName: \"kubernetes.io/projected/b6f07757-816a-4dd2-b804-795dd6a874fd-kube-api-access-v7knm\") pod \"collect-profiles-29533770-sjpgt\" (UID: \"b6f07757-816a-4dd2-b804-795dd6a874fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.380368 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6f07757-816a-4dd2-b804-795dd6a874fd-secret-volume\") pod \"collect-profiles-29533770-sjpgt\" (UID: \"b6f07757-816a-4dd2-b804-795dd6a874fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.380433 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6f07757-816a-4dd2-b804-795dd6a874fd-config-volume\") pod \"collect-profiles-29533770-sjpgt\" (UID: \"b6f07757-816a-4dd2-b804-795dd6a874fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.381888 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6f07757-816a-4dd2-b804-795dd6a874fd-config-volume\") pod \"collect-profiles-29533770-sjpgt\" (UID: \"b6f07757-816a-4dd2-b804-795dd6a874fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.387429 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6f07757-816a-4dd2-b804-795dd6a874fd-secret-volume\") pod \"collect-profiles-29533770-sjpgt\" (UID: \"b6f07757-816a-4dd2-b804-795dd6a874fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.398085 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7knm\" (UniqueName: \"kubernetes.io/projected/b6f07757-816a-4dd2-b804-795dd6a874fd-kube-api-access-v7knm\") pod \"collect-profiles-29533770-sjpgt\" (UID: \"b6f07757-816a-4dd2-b804-795dd6a874fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.409413 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsrkz\" (UniqueName: \"kubernetes.io/projected/67c273aa-5ff2-43fc-871f-64caafce5463-kube-api-access-xsrkz\") pod \"auto-csr-approver-29533770-qrbhh\" (UID: \"67c273aa-5ff2-43fc-871f-64caafce5463\") " pod="openshift-infra/auto-csr-approver-29533770-qrbhh" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.478209 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533770-qrbhh" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.497401 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt" Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.696290 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt"] Feb 25 13:30:00 crc kubenswrapper[4815]: W0225 13:30:00.708784 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6f07757_816a_4dd2_b804_795dd6a874fd.slice/crio-826fbba75e0ea35ce8bc9512a1551f8454451b9a9558f3bee3711305390b8b48 WatchSource:0}: Error finding container 826fbba75e0ea35ce8bc9512a1551f8454451b9a9558f3bee3711305390b8b48: Status 404 returned error can't find the container with id 826fbba75e0ea35ce8bc9512a1551f8454451b9a9558f3bee3711305390b8b48 Feb 25 13:30:00 crc kubenswrapper[4815]: I0225 13:30:00.740682 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533770-qrbhh"] Feb 25 13:30:00 crc kubenswrapper[4815]: W0225 13:30:00.744341 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67c273aa_5ff2_43fc_871f_64caafce5463.slice/crio-fc887198d971b8cbf3a8b9606378d51a8db465b4a4b3dafd2f6bb57a50d449a3 WatchSource:0}: Error finding container fc887198d971b8cbf3a8b9606378d51a8db465b4a4b3dafd2f6bb57a50d449a3: Status 404 returned error can't find the container with id fc887198d971b8cbf3a8b9606378d51a8db465b4a4b3dafd2f6bb57a50d449a3 Feb 25 13:30:01 crc kubenswrapper[4815]: I0225 13:30:01.507553 4815 generic.go:334] "Generic (PLEG): container finished" podID="b6f07757-816a-4dd2-b804-795dd6a874fd" containerID="cc3f45f46c471f80b48f0128822f6de4c029341b91547818a21a8764a1ad026a" exitCode=0 Feb 25 13:30:01 crc kubenswrapper[4815]: I0225 13:30:01.507705 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt" event={"ID":"b6f07757-816a-4dd2-b804-795dd6a874fd","Type":"ContainerDied","Data":"cc3f45f46c471f80b48f0128822f6de4c029341b91547818a21a8764a1ad026a"} Feb 25 13:30:01 crc kubenswrapper[4815]: I0225 13:30:01.507908 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt" event={"ID":"b6f07757-816a-4dd2-b804-795dd6a874fd","Type":"ContainerStarted","Data":"826fbba75e0ea35ce8bc9512a1551f8454451b9a9558f3bee3711305390b8b48"} Feb 25 13:30:01 crc kubenswrapper[4815]: I0225 13:30:01.509639 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533770-qrbhh" event={"ID":"67c273aa-5ff2-43fc-871f-64caafce5463","Type":"ContainerStarted","Data":"fc887198d971b8cbf3a8b9606378d51a8db465b4a4b3dafd2f6bb57a50d449a3"} Feb 25 13:30:02 crc kubenswrapper[4815]: I0225 13:30:02.519853 4815 generic.go:334] "Generic (PLEG): container finished" podID="67c273aa-5ff2-43fc-871f-64caafce5463" containerID="43314d9c57872caad1ed75670b9302c7293fc3dd8c04ba73a133bfa03c431ed3" exitCode=0 Feb 25 13:30:02 crc kubenswrapper[4815]: I0225 13:30:02.520331 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533770-qrbhh" event={"ID":"67c273aa-5ff2-43fc-871f-64caafce5463","Type":"ContainerDied","Data":"43314d9c57872caad1ed75670b9302c7293fc3dd8c04ba73a133bfa03c431ed3"} Feb 25 13:30:02 crc kubenswrapper[4815]: I0225 13:30:02.822257 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt" Feb 25 13:30:02 crc kubenswrapper[4815]: I0225 13:30:02.909495 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7knm\" (UniqueName: \"kubernetes.io/projected/b6f07757-816a-4dd2-b804-795dd6a874fd-kube-api-access-v7knm\") pod \"b6f07757-816a-4dd2-b804-795dd6a874fd\" (UID: \"b6f07757-816a-4dd2-b804-795dd6a874fd\") " Feb 25 13:30:02 crc kubenswrapper[4815]: I0225 13:30:02.909629 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6f07757-816a-4dd2-b804-795dd6a874fd-secret-volume\") pod \"b6f07757-816a-4dd2-b804-795dd6a874fd\" (UID: \"b6f07757-816a-4dd2-b804-795dd6a874fd\") " Feb 25 13:30:02 crc kubenswrapper[4815]: I0225 13:30:02.909669 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6f07757-816a-4dd2-b804-795dd6a874fd-config-volume\") pod \"b6f07757-816a-4dd2-b804-795dd6a874fd\" (UID: \"b6f07757-816a-4dd2-b804-795dd6a874fd\") " Feb 25 13:30:02 crc kubenswrapper[4815]: I0225 13:30:02.910714 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6f07757-816a-4dd2-b804-795dd6a874fd-config-volume" (OuterVolumeSpecName: "config-volume") pod "b6f07757-816a-4dd2-b804-795dd6a874fd" (UID: "b6f07757-816a-4dd2-b804-795dd6a874fd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:30:02 crc kubenswrapper[4815]: I0225 13:30:02.916679 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6f07757-816a-4dd2-b804-795dd6a874fd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b6f07757-816a-4dd2-b804-795dd6a874fd" (UID: "b6f07757-816a-4dd2-b804-795dd6a874fd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:30:02 crc kubenswrapper[4815]: I0225 13:30:02.917281 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6f07757-816a-4dd2-b804-795dd6a874fd-kube-api-access-v7knm" (OuterVolumeSpecName: "kube-api-access-v7knm") pod "b6f07757-816a-4dd2-b804-795dd6a874fd" (UID: "b6f07757-816a-4dd2-b804-795dd6a874fd"). InnerVolumeSpecName "kube-api-access-v7knm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:30:03 crc kubenswrapper[4815]: I0225 13:30:03.010683 4815 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6f07757-816a-4dd2-b804-795dd6a874fd-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:03 crc kubenswrapper[4815]: I0225 13:30:03.010731 4815 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6f07757-816a-4dd2-b804-795dd6a874fd-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:03 crc kubenswrapper[4815]: I0225 13:30:03.010865 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7knm\" (UniqueName: \"kubernetes.io/projected/b6f07757-816a-4dd2-b804-795dd6a874fd-kube-api-access-v7knm\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:03 crc kubenswrapper[4815]: I0225 13:30:03.530163 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt" Feb 25 13:30:03 crc kubenswrapper[4815]: I0225 13:30:03.531164 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt" event={"ID":"b6f07757-816a-4dd2-b804-795dd6a874fd","Type":"ContainerDied","Data":"826fbba75e0ea35ce8bc9512a1551f8454451b9a9558f3bee3711305390b8b48"} Feb 25 13:30:03 crc kubenswrapper[4815]: I0225 13:30:03.531251 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="826fbba75e0ea35ce8bc9512a1551f8454451b9a9558f3bee3711305390b8b48" Feb 25 13:30:03 crc kubenswrapper[4815]: I0225 13:30:03.842752 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533770-qrbhh" Feb 25 13:30:03 crc kubenswrapper[4815]: I0225 13:30:03.924199 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsrkz\" (UniqueName: \"kubernetes.io/projected/67c273aa-5ff2-43fc-871f-64caafce5463-kube-api-access-xsrkz\") pod \"67c273aa-5ff2-43fc-871f-64caafce5463\" (UID: \"67c273aa-5ff2-43fc-871f-64caafce5463\") " Feb 25 13:30:03 crc kubenswrapper[4815]: I0225 13:30:03.930002 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67c273aa-5ff2-43fc-871f-64caafce5463-kube-api-access-xsrkz" (OuterVolumeSpecName: "kube-api-access-xsrkz") pod "67c273aa-5ff2-43fc-871f-64caafce5463" (UID: "67c273aa-5ff2-43fc-871f-64caafce5463"). InnerVolumeSpecName "kube-api-access-xsrkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:30:04 crc kubenswrapper[4815]: I0225 13:30:04.026231 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsrkz\" (UniqueName: \"kubernetes.io/projected/67c273aa-5ff2-43fc-871f-64caafce5463-kube-api-access-xsrkz\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:04 crc kubenswrapper[4815]: I0225 13:30:04.538988 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533770-qrbhh" event={"ID":"67c273aa-5ff2-43fc-871f-64caafce5463","Type":"ContainerDied","Data":"fc887198d971b8cbf3a8b9606378d51a8db465b4a4b3dafd2f6bb57a50d449a3"} Feb 25 13:30:04 crc kubenswrapper[4815]: I0225 13:30:04.539031 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc887198d971b8cbf3a8b9606378d51a8db465b4a4b3dafd2f6bb57a50d449a3" Feb 25 13:30:04 crc kubenswrapper[4815]: I0225 13:30:04.539071 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533770-qrbhh" Feb 25 13:30:04 crc kubenswrapper[4815]: I0225 13:30:04.917869 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533764-9mvkb"] Feb 25 13:30:04 crc kubenswrapper[4815]: I0225 13:30:04.925101 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533764-9mvkb"] Feb 25 13:30:04 crc kubenswrapper[4815]: I0225 13:30:04.949336 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba102d89-25ea-40ad-ae0b-78746dbf6d2a" path="/var/lib/kubelet/pods/ba102d89-25ea-40ad-ae0b-78746dbf6d2a/volumes" Feb 25 13:30:21 crc kubenswrapper[4815]: I0225 13:30:21.342034 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:30:21 crc kubenswrapper[4815]: I0225 13:30:21.344504 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:30:43 crc kubenswrapper[4815]: I0225 13:30:43.388431 4815 scope.go:117] "RemoveContainer" containerID="ef223c7f15d02f5311304de3f4d229a0c93eba794dc87c41fa03c34388bf23c1" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.331407 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-24znl"] Feb 25 13:30:44 crc kubenswrapper[4815]: E0225 13:30:44.331960 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6f07757-816a-4dd2-b804-795dd6a874fd" containerName="collect-profiles" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.331984 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6f07757-816a-4dd2-b804-795dd6a874fd" containerName="collect-profiles" Feb 25 13:30:44 crc kubenswrapper[4815]: E0225 13:30:44.332006 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67c273aa-5ff2-43fc-871f-64caafce5463" containerName="oc" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.332015 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="67c273aa-5ff2-43fc-871f-64caafce5463" containerName="oc" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.332118 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="67c273aa-5ff2-43fc-871f-64caafce5463" containerName="oc" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.332131 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6f07757-816a-4dd2-b804-795dd6a874fd" containerName="collect-profiles" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.332464 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.359915 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-24znl"] Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.403950 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/61bd6796-9d05-404c-9a62-2f2970e772b1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.404003 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmzf7\" (UniqueName: \"kubernetes.io/projected/61bd6796-9d05-404c-9a62-2f2970e772b1-kube-api-access-dmzf7\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.404031 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61bd6796-9d05-404c-9a62-2f2970e772b1-bound-sa-token\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.404169 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/61bd6796-9d05-404c-9a62-2f2970e772b1-registry-tls\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.404225 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/61bd6796-9d05-404c-9a62-2f2970e772b1-registry-certificates\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.404266 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.404310 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/61bd6796-9d05-404c-9a62-2f2970e772b1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.404337 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61bd6796-9d05-404c-9a62-2f2970e772b1-trusted-ca\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.427230 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.505845 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmzf7\" (UniqueName: \"kubernetes.io/projected/61bd6796-9d05-404c-9a62-2f2970e772b1-kube-api-access-dmzf7\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.505900 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61bd6796-9d05-404c-9a62-2f2970e772b1-bound-sa-token\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.505919 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/61bd6796-9d05-404c-9a62-2f2970e772b1-registry-tls\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.505934 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/61bd6796-9d05-404c-9a62-2f2970e772b1-registry-certificates\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.505957 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/61bd6796-9d05-404c-9a62-2f2970e772b1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.505970 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61bd6796-9d05-404c-9a62-2f2970e772b1-trusted-ca\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.506009 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/61bd6796-9d05-404c-9a62-2f2970e772b1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.507224 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/61bd6796-9d05-404c-9a62-2f2970e772b1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.507975 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61bd6796-9d05-404c-9a62-2f2970e772b1-trusted-ca\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.508442 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/61bd6796-9d05-404c-9a62-2f2970e772b1-registry-certificates\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.512111 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/61bd6796-9d05-404c-9a62-2f2970e772b1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.512142 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/61bd6796-9d05-404c-9a62-2f2970e772b1-registry-tls\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.523176 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmzf7\" (UniqueName: \"kubernetes.io/projected/61bd6796-9d05-404c-9a62-2f2970e772b1-kube-api-access-dmzf7\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.529394 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61bd6796-9d05-404c-9a62-2f2970e772b1-bound-sa-token\") pod \"image-registry-66df7c8f76-24znl\" (UID: \"61bd6796-9d05-404c-9a62-2f2970e772b1\") " pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.644884 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.856736 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-hxclh"] Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.857687 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-hxclh" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.861560 4815 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-wc725" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.861767 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.861929 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.867842 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-hxclh"] Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.876690 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-xf9wz"] Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.877317 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-xf9wz" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.879424 4815 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-ld654" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.882479 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-qslns"] Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.883239 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-qslns" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.885522 4815 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-hprzb" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.912006 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-xf9wz"] Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.912600 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2d96\" (UniqueName: \"kubernetes.io/projected/ecdcd2c0-4b94-44aa-91fe-0fd1b09909f1-kube-api-access-f2d96\") pod \"cert-manager-cainjector-cf98fcc89-hxclh\" (UID: \"ecdcd2c0-4b94-44aa-91fe-0fd1b09909f1\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-hxclh" Feb 25 13:30:44 crc kubenswrapper[4815]: I0225 13:30:44.918151 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-qslns"] Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.014072 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrxvs\" (UniqueName: \"kubernetes.io/projected/7b298115-612a-4691-8e63-067ce1230fdf-kube-api-access-rrxvs\") pod \"cert-manager-webhook-687f57d79b-qslns\" (UID: \"7b298115-612a-4691-8e63-067ce1230fdf\") " pod="cert-manager/cert-manager-webhook-687f57d79b-qslns" Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.014143 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2d96\" (UniqueName: \"kubernetes.io/projected/ecdcd2c0-4b94-44aa-91fe-0fd1b09909f1-kube-api-access-f2d96\") pod \"cert-manager-cainjector-cf98fcc89-hxclh\" (UID: \"ecdcd2c0-4b94-44aa-91fe-0fd1b09909f1\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-hxclh" Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.014222 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwvjw\" (UniqueName: \"kubernetes.io/projected/bc773872-e8b3-4d04-acb3-593cb469b261-kube-api-access-jwvjw\") pod \"cert-manager-858654f9db-xf9wz\" (UID: \"bc773872-e8b3-4d04-acb3-593cb469b261\") " pod="cert-manager/cert-manager-858654f9db-xf9wz" Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.033003 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2d96\" (UniqueName: \"kubernetes.io/projected/ecdcd2c0-4b94-44aa-91fe-0fd1b09909f1-kube-api-access-f2d96\") pod \"cert-manager-cainjector-cf98fcc89-hxclh\" (UID: \"ecdcd2c0-4b94-44aa-91fe-0fd1b09909f1\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-hxclh" Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.116883 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrxvs\" (UniqueName: \"kubernetes.io/projected/7b298115-612a-4691-8e63-067ce1230fdf-kube-api-access-rrxvs\") pod \"cert-manager-webhook-687f57d79b-qslns\" (UID: \"7b298115-612a-4691-8e63-067ce1230fdf\") " pod="cert-manager/cert-manager-webhook-687f57d79b-qslns" Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.116990 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwvjw\" (UniqueName: \"kubernetes.io/projected/bc773872-e8b3-4d04-acb3-593cb469b261-kube-api-access-jwvjw\") pod \"cert-manager-858654f9db-xf9wz\" (UID: \"bc773872-e8b3-4d04-acb3-593cb469b261\") " pod="cert-manager/cert-manager-858654f9db-xf9wz" Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.132170 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrxvs\" (UniqueName: \"kubernetes.io/projected/7b298115-612a-4691-8e63-067ce1230fdf-kube-api-access-rrxvs\") pod \"cert-manager-webhook-687f57d79b-qslns\" (UID: \"7b298115-612a-4691-8e63-067ce1230fdf\") " pod="cert-manager/cert-manager-webhook-687f57d79b-qslns" Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.137050 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-24znl"] Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.137149 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwvjw\" (UniqueName: \"kubernetes.io/projected/bc773872-e8b3-4d04-acb3-593cb469b261-kube-api-access-jwvjw\") pod \"cert-manager-858654f9db-xf9wz\" (UID: \"bc773872-e8b3-4d04-acb3-593cb469b261\") " pod="cert-manager/cert-manager-858654f9db-xf9wz" Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.176696 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-hxclh" Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.199849 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-xf9wz" Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.205422 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-qslns" Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.513586 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-xf9wz"] Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.623436 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-hxclh"] Feb 25 13:30:45 crc kubenswrapper[4815]: W0225 13:30:45.628128 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecdcd2c0_4b94_44aa_91fe_0fd1b09909f1.slice/crio-083088db5e8329f529de16a196bf09ccd845809a42b84e8997fc22b7c5efb7cd WatchSource:0}: Error finding container 083088db5e8329f529de16a196bf09ccd845809a42b84e8997fc22b7c5efb7cd: Status 404 returned error can't find the container with id 083088db5e8329f529de16a196bf09ccd845809a42b84e8997fc22b7c5efb7cd Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.664330 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-qslns"] Feb 25 13:30:45 crc kubenswrapper[4815]: W0225 13:30:45.666184 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b298115_612a_4691_8e63_067ce1230fdf.slice/crio-a59f2e9845dfe0057f4846c354efd0c85e093ed7a9e37eab8aa08da250535c93 WatchSource:0}: Error finding container a59f2e9845dfe0057f4846c354efd0c85e093ed7a9e37eab8aa08da250535c93: Status 404 returned error can't find the container with id a59f2e9845dfe0057f4846c354efd0c85e093ed7a9e37eab8aa08da250535c93 Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.841059 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-hxclh" event={"ID":"ecdcd2c0-4b94-44aa-91fe-0fd1b09909f1","Type":"ContainerStarted","Data":"083088db5e8329f529de16a196bf09ccd845809a42b84e8997fc22b7c5efb7cd"} Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.843028 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-24znl" event={"ID":"61bd6796-9d05-404c-9a62-2f2970e772b1","Type":"ContainerStarted","Data":"344a1f256b7cb77c062e524bbd8bcade6ea8c99ae3f7d4a7edf04fe2c8f01c12"} Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.843059 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-24znl" event={"ID":"61bd6796-9d05-404c-9a62-2f2970e772b1","Type":"ContainerStarted","Data":"fa114143144878242bb4213cd7f9f461ccfa6088e99d9a99cc3f522752605d1f"} Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.843956 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.844970 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-qslns" event={"ID":"7b298115-612a-4691-8e63-067ce1230fdf","Type":"ContainerStarted","Data":"a59f2e9845dfe0057f4846c354efd0c85e093ed7a9e37eab8aa08da250535c93"} Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.847054 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-xf9wz" event={"ID":"bc773872-e8b3-4d04-acb3-593cb469b261","Type":"ContainerStarted","Data":"a9c021da42a32f5462fd5ea251d5c11ac1ac40d00ca6cd9e93beb1b090239ee8"} Feb 25 13:30:45 crc kubenswrapper[4815]: I0225 13:30:45.869603 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-24znl" podStartSLOduration=1.869559556 podStartE2EDuration="1.869559556s" podCreationTimestamp="2026-02-25 13:30:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:30:45.865524554 +0000 UTC m=+603.666622618" watchObservedRunningTime="2026-02-25 13:30:45.869559556 +0000 UTC m=+603.670657620" Feb 25 13:30:49 crc kubenswrapper[4815]: I0225 13:30:49.891760 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-hxclh" event={"ID":"ecdcd2c0-4b94-44aa-91fe-0fd1b09909f1","Type":"ContainerStarted","Data":"b21dc6bea966e87e5f98301f44918bef89bb3a5b12991e0a5871759ee7a673ee"} Feb 25 13:30:49 crc kubenswrapper[4815]: I0225 13:30:49.897500 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-qslns" event={"ID":"7b298115-612a-4691-8e63-067ce1230fdf","Type":"ContainerStarted","Data":"e629eef3d29f88e970318fc61f1733d815305dc63d6657465c2e1124c1109213"} Feb 25 13:30:49 crc kubenswrapper[4815]: I0225 13:30:49.897643 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-qslns" Feb 25 13:30:49 crc kubenswrapper[4815]: I0225 13:30:49.900108 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-xf9wz" event={"ID":"bc773872-e8b3-4d04-acb3-593cb469b261","Type":"ContainerStarted","Data":"37f5ee2ca684a8268dc6a1ed913f7de859230ec1ed1512a0480b7849ab571f93"} Feb 25 13:30:49 crc kubenswrapper[4815]: I0225 13:30:49.950495 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-hxclh" podStartSLOduration=2.479436814 podStartE2EDuration="5.950479634s" podCreationTimestamp="2026-02-25 13:30:44 +0000 UTC" firstStartedPulling="2026-02-25 13:30:45.63026223 +0000 UTC m=+603.431360284" lastFinishedPulling="2026-02-25 13:30:49.10130504 +0000 UTC m=+606.902403104" observedRunningTime="2026-02-25 13:30:49.949236843 +0000 UTC m=+607.750334897" watchObservedRunningTime="2026-02-25 13:30:49.950479634 +0000 UTC m=+607.751577698" Feb 25 13:30:49 crc kubenswrapper[4815]: I0225 13:30:49.968404 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-xf9wz" podStartSLOduration=2.392195454 podStartE2EDuration="5.968387295s" podCreationTimestamp="2026-02-25 13:30:44 +0000 UTC" firstStartedPulling="2026-02-25 13:30:45.528641813 +0000 UTC m=+603.329739867" lastFinishedPulling="2026-02-25 13:30:49.104833644 +0000 UTC m=+606.905931708" observedRunningTime="2026-02-25 13:30:49.967841718 +0000 UTC m=+607.768939832" watchObservedRunningTime="2026-02-25 13:30:49.968387295 +0000 UTC m=+607.769485349" Feb 25 13:30:49 crc kubenswrapper[4815]: I0225 13:30:49.997213 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-qslns" podStartSLOduration=2.554354605 podStartE2EDuration="5.997181279s" podCreationTimestamp="2026-02-25 13:30:44 +0000 UTC" firstStartedPulling="2026-02-25 13:30:45.668532782 +0000 UTC m=+603.469630836" lastFinishedPulling="2026-02-25 13:30:49.111359446 +0000 UTC m=+606.912457510" observedRunningTime="2026-02-25 13:30:49.990978698 +0000 UTC m=+607.792076752" watchObservedRunningTime="2026-02-25 13:30:49.997181279 +0000 UTC m=+607.798279353" Feb 25 13:30:51 crc kubenswrapper[4815]: I0225 13:30:51.342443 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:30:51 crc kubenswrapper[4815]: I0225 13:30:51.342556 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.803133 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qkfgf"] Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.804051 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="ovn-controller" containerID="cri-o://8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90" gracePeriod=30 Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.804087 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="nbdb" containerID="cri-o://80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285" gracePeriod=30 Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.804200 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="northd" containerID="cri-o://e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0" gracePeriod=30 Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.804271 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820" gracePeriod=30 Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.804341 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="kube-rbac-proxy-node" containerID="cri-o://b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4" gracePeriod=30 Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.804398 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="ovn-acl-logging" containerID="cri-o://9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743" gracePeriod=30 Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.804665 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="sbdb" containerID="cri-o://022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4" gracePeriod=30 Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.847726 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="ovnkube-controller" containerID="cri-o://d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837" gracePeriod=30 Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.950762 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkfgf_6cd6cca6-d1e9-4c56-8f4d-f396a769ba07/ovn-acl-logging/0.log" Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.951274 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkfgf_6cd6cca6-d1e9-4c56-8f4d-f396a769ba07/ovn-controller/0.log" Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.952218 4815 generic.go:334] "Generic (PLEG): container finished" podID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerID="ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820" exitCode=0 Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.952323 4815 generic.go:334] "Generic (PLEG): container finished" podID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerID="b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4" exitCode=0 Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.952405 4815 generic.go:334] "Generic (PLEG): container finished" podID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerID="9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743" exitCode=143 Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.952425 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerDied","Data":"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820"} Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.952564 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerDied","Data":"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4"} Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.952582 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerDied","Data":"9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743"} Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.957259 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-l92cc_88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6/kube-multus/0.log" Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.957300 4815 generic.go:334] "Generic (PLEG): container finished" podID="88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6" containerID="86c30dd3c376e2edfbfc035c5f37c6d85c659e8fe429f8c162ca3e14df9048c5" exitCode=2 Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.957323 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-l92cc" event={"ID":"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6","Type":"ContainerDied","Data":"86c30dd3c376e2edfbfc035c5f37c6d85c659e8fe429f8c162ca3e14df9048c5"} Feb 25 13:30:54 crc kubenswrapper[4815]: I0225 13:30:54.957720 4815 scope.go:117] "RemoveContainer" containerID="86c30dd3c376e2edfbfc035c5f37c6d85c659e8fe429f8c162ca3e14df9048c5" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.154095 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkfgf_6cd6cca6-d1e9-4c56-8f4d-f396a769ba07/ovn-acl-logging/0.log" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.154926 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkfgf_6cd6cca6-d1e9-4c56-8f4d-f396a769ba07/ovn-controller/0.log" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.155361 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.219263 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-qslns" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.221778 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qc8ls"] Feb 25 13:30:55 crc kubenswrapper[4815]: E0225 13:30:55.222086 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="kubecfg-setup" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.222118 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="kubecfg-setup" Feb 25 13:30:55 crc kubenswrapper[4815]: E0225 13:30:55.222139 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="nbdb" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.222150 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="nbdb" Feb 25 13:30:55 crc kubenswrapper[4815]: E0225 13:30:55.222162 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="kube-rbac-proxy-ovn-metrics" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.222172 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="kube-rbac-proxy-ovn-metrics" Feb 25 13:30:55 crc kubenswrapper[4815]: E0225 13:30:55.222187 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="sbdb" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.222196 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="sbdb" Feb 25 13:30:55 crc kubenswrapper[4815]: E0225 13:30:55.222210 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="ovn-acl-logging" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.222220 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="ovn-acl-logging" Feb 25 13:30:55 crc kubenswrapper[4815]: E0225 13:30:55.222239 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="kube-rbac-proxy-node" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.222250 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="kube-rbac-proxy-node" Feb 25 13:30:55 crc kubenswrapper[4815]: E0225 13:30:55.222269 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="northd" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.222282 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="northd" Feb 25 13:30:55 crc kubenswrapper[4815]: E0225 13:30:55.222293 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="ovnkube-controller" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.222304 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="ovnkube-controller" Feb 25 13:30:55 crc kubenswrapper[4815]: E0225 13:30:55.222320 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="ovn-controller" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.222330 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="ovn-controller" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.222461 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="kube-rbac-proxy-node" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.222475 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="ovn-acl-logging" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.222484 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="ovn-controller" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.222496 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="sbdb" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.222537 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="nbdb" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.222550 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="northd" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.222566 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="kube-rbac-proxy-ovn-metrics" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.222578 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerName="ovnkube-controller" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.225637 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290039 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-var-lib-openvswitch\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290093 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-env-overrides\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290127 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-etc-openvswitch\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290159 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovn-node-metrics-cert\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290189 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovnkube-script-lib\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290229 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-cni-bin\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290254 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-var-lib-cni-networks-ovn-kubernetes\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290276 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-systemd-units\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290303 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-run-ovn-kubernetes\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290320 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-systemd\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290344 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-ovn\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290180 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290368 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovnkube-config\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290295 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290347 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290422 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290421 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290390 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-cni-netd\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290444 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290481 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290503 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-kubelet\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290566 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-log-socket\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290573 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290588 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-run-netns\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290613 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290632 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-openvswitch\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290639 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-log-socket" (OuterVolumeSpecName: "log-socket") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290656 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-node-log\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290694 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbpb8\" (UniqueName: \"kubernetes.io/projected/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-kube-api-access-cbpb8\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290724 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-slash\") pod \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\" (UID: \"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07\") " Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290660 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290678 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290698 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290689 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290733 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-node-log" (OuterVolumeSpecName: "node-log") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.290983 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291016 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-slash" (OuterVolumeSpecName: "host-slash") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291139 4815 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-slash\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291156 4815 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291171 4815 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291183 4815 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291195 4815 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291207 4815 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291220 4815 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291232 4815 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291244 4815 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291256 4815 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291267 4815 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291280 4815 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291291 4815 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291301 4815 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-log-socket\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291313 4815 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291325 4815 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-node-log\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.291336 4815 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.295915 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-kube-api-access-cbpb8" (OuterVolumeSpecName: "kube-api-access-cbpb8") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "kube-api-access-cbpb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.297745 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.313489 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" (UID: "6cd6cca6-d1e9-4c56-8f4d-f396a769ba07"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.392727 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-run-netns\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.392787 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-ovn-node-metrics-cert\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.392817 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-etc-openvswitch\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.392841 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.392872 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-run-ovn-kubernetes\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.392894 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-run-ovn\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.392923 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-ovnkube-script-lib\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.392956 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-cni-netd\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.392978 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z482\" (UniqueName: \"kubernetes.io/projected/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-kube-api-access-5z482\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.392997 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-cni-bin\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.393031 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-kubelet\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.393056 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-log-socket\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.393080 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-env-overrides\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.393117 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-slash\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.393138 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-run-systemd\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.393156 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-node-log\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.393179 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-var-lib-openvswitch\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.393202 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-systemd-units\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.393226 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-ovnkube-config\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.393248 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-run-openvswitch\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.393296 4815 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.393311 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbpb8\" (UniqueName: \"kubernetes.io/projected/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-kube-api-access-cbpb8\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.393325 4815 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.494847 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-run-netns\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.494907 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-ovn-node-metrics-cert\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.494942 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.494972 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-etc-openvswitch\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495015 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-run-ovn\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495040 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-run-ovn-kubernetes\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495079 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-ovnkube-script-lib\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495126 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-cni-netd\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495155 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z482\" (UniqueName: \"kubernetes.io/projected/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-kube-api-access-5z482\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495182 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-cni-bin\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495233 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-kubelet\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495264 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-log-socket\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495292 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-env-overrides\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495324 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-slash\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495353 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-run-systemd\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495378 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-node-log\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495407 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-var-lib-openvswitch\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495437 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-systemd-units\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495470 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-ovnkube-config\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495498 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-run-openvswitch\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495643 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-run-openvswitch\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.495703 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-run-netns\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.496288 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-kubelet\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.496367 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-run-ovn-kubernetes\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.496433 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.496461 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-etc-openvswitch\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.496483 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-run-ovn\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.496524 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-var-lib-openvswitch\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.496547 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-node-log\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.496830 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-cni-netd\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.496905 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-systemd-units\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.496929 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-cni-bin\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.497321 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-log-socket\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.497363 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-ovnkube-script-lib\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.497377 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-env-overrides\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.497409 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-host-slash\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.497844 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-ovnkube-config\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.496319 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-run-systemd\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.500208 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-ovn-node-metrics-cert\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.525237 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z482\" (UniqueName: \"kubernetes.io/projected/bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8-kube-api-access-5z482\") pod \"ovnkube-node-qc8ls\" (UID: \"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.546551 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:30:55 crc kubenswrapper[4815]: W0225 13:30:55.563344 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd137b1b_e3bd_4176_8ceb_86ac1faeb2e8.slice/crio-03cbcc2375144757d17cea1328082ee8db8015577eedcedc0126c2abde0a16c9 WatchSource:0}: Error finding container 03cbcc2375144757d17cea1328082ee8db8015577eedcedc0126c2abde0a16c9: Status 404 returned error can't find the container with id 03cbcc2375144757d17cea1328082ee8db8015577eedcedc0126c2abde0a16c9 Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.974322 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-l92cc_88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6/kube-multus/0.log" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.974937 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-l92cc" event={"ID":"88b40c3e-14f9-4ec2-bd13-2fecf33ec9e6","Type":"ContainerStarted","Data":"d932d34a310c54c9bdb0b79900e925a3494f7fd6d926d0570646dbec0e0244d1"} Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.983469 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkfgf_6cd6cca6-d1e9-4c56-8f4d-f396a769ba07/ovn-acl-logging/0.log" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.984483 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qkfgf_6cd6cca6-d1e9-4c56-8f4d-f396a769ba07/ovn-controller/0.log" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.985215 4815 generic.go:334] "Generic (PLEG): container finished" podID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerID="d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837" exitCode=0 Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.985259 4815 generic.go:334] "Generic (PLEG): container finished" podID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerID="022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4" exitCode=0 Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.985277 4815 generic.go:334] "Generic (PLEG): container finished" podID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerID="80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285" exitCode=0 Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.985292 4815 generic.go:334] "Generic (PLEG): container finished" podID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerID="e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0" exitCode=0 Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.985312 4815 generic.go:334] "Generic (PLEG): container finished" podID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" containerID="8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90" exitCode=143 Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.985332 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.985397 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerDied","Data":"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837"} Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.985441 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerDied","Data":"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4"} Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.985465 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerDied","Data":"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285"} Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.985485 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerDied","Data":"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0"} Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.985535 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerDied","Data":"8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90"} Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.985557 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qkfgf" event={"ID":"6cd6cca6-d1e9-4c56-8f4d-f396a769ba07","Type":"ContainerDied","Data":"0d7781213d2e22dc8880868392952084c00d31ea6d52ec3244161ae3b53ca60e"} Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.985577 4815 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743"} Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.985594 4815 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90"} Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.985606 4815 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085"} Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.985630 4815 scope.go:117] "RemoveContainer" containerID="d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837" Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.989783 4815 generic.go:334] "Generic (PLEG): container finished" podID="bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8" containerID="2364e3e966a23a86e3df598c08b06885484dae041f6052b4287b71c1e33b1fe7" exitCode=0 Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.989826 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" event={"ID":"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8","Type":"ContainerDied","Data":"2364e3e966a23a86e3df598c08b06885484dae041f6052b4287b71c1e33b1fe7"} Feb 25 13:30:55 crc kubenswrapper[4815]: I0225 13:30:55.989855 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" event={"ID":"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8","Type":"ContainerStarted","Data":"03cbcc2375144757d17cea1328082ee8db8015577eedcedc0126c2abde0a16c9"} Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.025864 4815 scope.go:117] "RemoveContainer" containerID="022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.058676 4815 scope.go:117] "RemoveContainer" containerID="80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.094329 4815 scope.go:117] "RemoveContainer" containerID="e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.098908 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qkfgf"] Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.105936 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qkfgf"] Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.117205 4815 scope.go:117] "RemoveContainer" containerID="ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.136850 4815 scope.go:117] "RemoveContainer" containerID="b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.152545 4815 scope.go:117] "RemoveContainer" containerID="9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.165253 4815 scope.go:117] "RemoveContainer" containerID="8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.187009 4815 scope.go:117] "RemoveContainer" containerID="ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.213215 4815 scope.go:117] "RemoveContainer" containerID="d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837" Feb 25 13:30:56 crc kubenswrapper[4815]: E0225 13:30:56.213661 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837\": container with ID starting with d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837 not found: ID does not exist" containerID="d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.213694 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837"} err="failed to get container status \"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837\": rpc error: code = NotFound desc = could not find container \"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837\": container with ID starting with d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.213721 4815 scope.go:117] "RemoveContainer" containerID="022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4" Feb 25 13:30:56 crc kubenswrapper[4815]: E0225 13:30:56.214023 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4\": container with ID starting with 022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4 not found: ID does not exist" containerID="022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.214049 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4"} err="failed to get container status \"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4\": rpc error: code = NotFound desc = could not find container \"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4\": container with ID starting with 022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.214063 4815 scope.go:117] "RemoveContainer" containerID="80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285" Feb 25 13:30:56 crc kubenswrapper[4815]: E0225 13:30:56.214319 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285\": container with ID starting with 80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285 not found: ID does not exist" containerID="80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.214341 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285"} err="failed to get container status \"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285\": rpc error: code = NotFound desc = could not find container \"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285\": container with ID starting with 80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.214352 4815 scope.go:117] "RemoveContainer" containerID="e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0" Feb 25 13:30:56 crc kubenswrapper[4815]: E0225 13:30:56.220021 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0\": container with ID starting with e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0 not found: ID does not exist" containerID="e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.220061 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0"} err="failed to get container status \"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0\": rpc error: code = NotFound desc = could not find container \"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0\": container with ID starting with e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.220083 4815 scope.go:117] "RemoveContainer" containerID="ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820" Feb 25 13:30:56 crc kubenswrapper[4815]: E0225 13:30:56.220555 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820\": container with ID starting with ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820 not found: ID does not exist" containerID="ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.220586 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820"} err="failed to get container status \"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820\": rpc error: code = NotFound desc = could not find container \"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820\": container with ID starting with ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.220603 4815 scope.go:117] "RemoveContainer" containerID="b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4" Feb 25 13:30:56 crc kubenswrapper[4815]: E0225 13:30:56.220952 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4\": container with ID starting with b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4 not found: ID does not exist" containerID="b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.220974 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4"} err="failed to get container status \"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4\": rpc error: code = NotFound desc = could not find container \"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4\": container with ID starting with b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.220987 4815 scope.go:117] "RemoveContainer" containerID="9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743" Feb 25 13:30:56 crc kubenswrapper[4815]: E0225 13:30:56.221167 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743\": container with ID starting with 9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743 not found: ID does not exist" containerID="9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.221189 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743"} err="failed to get container status \"9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743\": rpc error: code = NotFound desc = could not find container \"9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743\": container with ID starting with 9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.221202 4815 scope.go:117] "RemoveContainer" containerID="8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90" Feb 25 13:30:56 crc kubenswrapper[4815]: E0225 13:30:56.221438 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90\": container with ID starting with 8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90 not found: ID does not exist" containerID="8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.221469 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90"} err="failed to get container status \"8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90\": rpc error: code = NotFound desc = could not find container \"8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90\": container with ID starting with 8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.221488 4815 scope.go:117] "RemoveContainer" containerID="ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085" Feb 25 13:30:56 crc kubenswrapper[4815]: E0225 13:30:56.221809 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085\": container with ID starting with ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085 not found: ID does not exist" containerID="ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.221839 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085"} err="failed to get container status \"ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085\": rpc error: code = NotFound desc = could not find container \"ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085\": container with ID starting with ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.221857 4815 scope.go:117] "RemoveContainer" containerID="d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.222066 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837"} err="failed to get container status \"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837\": rpc error: code = NotFound desc = could not find container \"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837\": container with ID starting with d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.222091 4815 scope.go:117] "RemoveContainer" containerID="022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.226385 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4"} err="failed to get container status \"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4\": rpc error: code = NotFound desc = could not find container \"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4\": container with ID starting with 022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.226415 4815 scope.go:117] "RemoveContainer" containerID="80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.226708 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285"} err="failed to get container status \"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285\": rpc error: code = NotFound desc = could not find container \"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285\": container with ID starting with 80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.226731 4815 scope.go:117] "RemoveContainer" containerID="e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.226925 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0"} err="failed to get container status \"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0\": rpc error: code = NotFound desc = could not find container \"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0\": container with ID starting with e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.226951 4815 scope.go:117] "RemoveContainer" containerID="ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.227154 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820"} err="failed to get container status \"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820\": rpc error: code = NotFound desc = could not find container \"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820\": container with ID starting with ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.227185 4815 scope.go:117] "RemoveContainer" containerID="b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.227827 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4"} err="failed to get container status \"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4\": rpc error: code = NotFound desc = could not find container \"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4\": container with ID starting with b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.227854 4815 scope.go:117] "RemoveContainer" containerID="9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.228046 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743"} err="failed to get container status \"9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743\": rpc error: code = NotFound desc = could not find container \"9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743\": container with ID starting with 9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.228067 4815 scope.go:117] "RemoveContainer" containerID="8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.228224 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90"} err="failed to get container status \"8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90\": rpc error: code = NotFound desc = could not find container \"8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90\": container with ID starting with 8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.228241 4815 scope.go:117] "RemoveContainer" containerID="ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.228386 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085"} err="failed to get container status \"ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085\": rpc error: code = NotFound desc = could not find container \"ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085\": container with ID starting with ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.228403 4815 scope.go:117] "RemoveContainer" containerID="d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.228558 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837"} err="failed to get container status \"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837\": rpc error: code = NotFound desc = could not find container \"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837\": container with ID starting with d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.228575 4815 scope.go:117] "RemoveContainer" containerID="022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.228717 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4"} err="failed to get container status \"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4\": rpc error: code = NotFound desc = could not find container \"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4\": container with ID starting with 022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.228734 4815 scope.go:117] "RemoveContainer" containerID="80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.228878 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285"} err="failed to get container status \"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285\": rpc error: code = NotFound desc = could not find container \"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285\": container with ID starting with 80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.228896 4815 scope.go:117] "RemoveContainer" containerID="e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.229023 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0"} err="failed to get container status \"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0\": rpc error: code = NotFound desc = could not find container \"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0\": container with ID starting with e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.229039 4815 scope.go:117] "RemoveContainer" containerID="ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.229169 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820"} err="failed to get container status \"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820\": rpc error: code = NotFound desc = could not find container \"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820\": container with ID starting with ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.229185 4815 scope.go:117] "RemoveContainer" containerID="b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.229316 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4"} err="failed to get container status \"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4\": rpc error: code = NotFound desc = could not find container \"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4\": container with ID starting with b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.229333 4815 scope.go:117] "RemoveContainer" containerID="9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.229460 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743"} err="failed to get container status \"9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743\": rpc error: code = NotFound desc = could not find container \"9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743\": container with ID starting with 9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.229478 4815 scope.go:117] "RemoveContainer" containerID="8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.229634 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90"} err="failed to get container status \"8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90\": rpc error: code = NotFound desc = could not find container \"8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90\": container with ID starting with 8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.229652 4815 scope.go:117] "RemoveContainer" containerID="ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.229786 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085"} err="failed to get container status \"ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085\": rpc error: code = NotFound desc = could not find container \"ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085\": container with ID starting with ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.229802 4815 scope.go:117] "RemoveContainer" containerID="d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.229940 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837"} err="failed to get container status \"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837\": rpc error: code = NotFound desc = could not find container \"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837\": container with ID starting with d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.229958 4815 scope.go:117] "RemoveContainer" containerID="022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.230128 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4"} err="failed to get container status \"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4\": rpc error: code = NotFound desc = could not find container \"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4\": container with ID starting with 022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.230153 4815 scope.go:117] "RemoveContainer" containerID="80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.230341 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285"} err="failed to get container status \"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285\": rpc error: code = NotFound desc = could not find container \"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285\": container with ID starting with 80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.230370 4815 scope.go:117] "RemoveContainer" containerID="e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.230567 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0"} err="failed to get container status \"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0\": rpc error: code = NotFound desc = could not find container \"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0\": container with ID starting with e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.230586 4815 scope.go:117] "RemoveContainer" containerID="ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.230826 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820"} err="failed to get container status \"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820\": rpc error: code = NotFound desc = could not find container \"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820\": container with ID starting with ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.230844 4815 scope.go:117] "RemoveContainer" containerID="b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.230979 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4"} err="failed to get container status \"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4\": rpc error: code = NotFound desc = could not find container \"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4\": container with ID starting with b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.230994 4815 scope.go:117] "RemoveContainer" containerID="9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.231121 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743"} err="failed to get container status \"9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743\": rpc error: code = NotFound desc = could not find container \"9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743\": container with ID starting with 9e7949fc9d74c38a6fdb1b56d66895e37181e52e7fd1be0af0e02b2cfd56e743 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.231136 4815 scope.go:117] "RemoveContainer" containerID="8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.231267 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90"} err="failed to get container status \"8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90\": rpc error: code = NotFound desc = could not find container \"8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90\": container with ID starting with 8504db7efcdce5874f7c680ea7a78fee099335ccb372c69e66b198344c115a90 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.231283 4815 scope.go:117] "RemoveContainer" containerID="ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.231413 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085"} err="failed to get container status \"ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085\": rpc error: code = NotFound desc = could not find container \"ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085\": container with ID starting with ba1068ede15fecfc5fca39d537a227b19e46f8d33e819a6855714dbbb9853085 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.231428 4815 scope.go:117] "RemoveContainer" containerID="d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.231589 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837"} err="failed to get container status \"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837\": rpc error: code = NotFound desc = could not find container \"d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837\": container with ID starting with d5768e5a7a8e3971aec0aecc8aef6933a506a28479442a2b9b1e0a0096cb9837 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.231607 4815 scope.go:117] "RemoveContainer" containerID="022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.231758 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4"} err="failed to get container status \"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4\": rpc error: code = NotFound desc = could not find container \"022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4\": container with ID starting with 022d3b9ffa2fa762d9dc390b4b66320071d409937257cbdb07cf41b3934298c4 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.231774 4815 scope.go:117] "RemoveContainer" containerID="80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.231917 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285"} err="failed to get container status \"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285\": rpc error: code = NotFound desc = could not find container \"80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285\": container with ID starting with 80f3025eb1b0c1ed7bcb9b703ea33b03951bb3dd5290eacd1bcfd4b0950db285 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.231932 4815 scope.go:117] "RemoveContainer" containerID="e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.232105 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0"} err="failed to get container status \"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0\": rpc error: code = NotFound desc = could not find container \"e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0\": container with ID starting with e3748f7c244b33a5ac1daf5469099e6f6beea7ca80ccf07fdd7c1b00203aa5d0 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.232127 4815 scope.go:117] "RemoveContainer" containerID="ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.232285 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820"} err="failed to get container status \"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820\": rpc error: code = NotFound desc = could not find container \"ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820\": container with ID starting with ed906ad14b5b197fe040b974e108901f2748008f5640c79d76aa1cd235b12820 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.232297 4815 scope.go:117] "RemoveContainer" containerID="b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.232442 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4"} err="failed to get container status \"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4\": rpc error: code = NotFound desc = could not find container \"b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4\": container with ID starting with b03d55a0c19ee606913b89e49e40baa74a93e0eb6a0469dcd65205036d90d5a4 not found: ID does not exist" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.942780 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cd6cca6-d1e9-4c56-8f4d-f396a769ba07" path="/var/lib/kubelet/pods/6cd6cca6-d1e9-4c56-8f4d-f396a769ba07/volumes" Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.998440 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" event={"ID":"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8","Type":"ContainerStarted","Data":"c4d5e55914ad60ba7e98ad91bd77e8700e7b7ee9cd2e06d28e36470756051688"} Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.998477 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" event={"ID":"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8","Type":"ContainerStarted","Data":"dcdda2c0804ebf5f165ec237823b42f0c45bca243aa6f44a8e42b437d569ab86"} Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.998489 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" event={"ID":"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8","Type":"ContainerStarted","Data":"f1406957ff0b31ff5621d50e58098fc7de106938a5af053d88ea2b35d3340039"} Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.998498 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" event={"ID":"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8","Type":"ContainerStarted","Data":"93811c3d6ae4afa9188cc36e5994fecf7769ff1ee926d0aafa724744603be994"} Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.998537 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" event={"ID":"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8","Type":"ContainerStarted","Data":"1df14bec2268e04e786ce5b61368ade86226ecb83b83ecf19c911b6027a0d859"} Feb 25 13:30:56 crc kubenswrapper[4815]: I0225 13:30:56.998546 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" event={"ID":"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8","Type":"ContainerStarted","Data":"b9607b3e46c850bf302d9e4e2c85f68f6e77652b6212ca56b536eaf30b439a9f"} Feb 25 13:31:00 crc kubenswrapper[4815]: I0225 13:31:00.026018 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" event={"ID":"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8","Type":"ContainerStarted","Data":"b4f5dfa6c7c1ebdb6069718c3dc0b40a51e614cb177d40962c1659897f6d35de"} Feb 25 13:31:02 crc kubenswrapper[4815]: I0225 13:31:02.044180 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" event={"ID":"bd137b1b-e3bd-4176-8ceb-86ac1faeb2e8","Type":"ContainerStarted","Data":"a4a5da9d18492f685f54850f0c6823a735e053dc01fe5dc9cc3ebab7e8f4a68d"} Feb 25 13:31:02 crc kubenswrapper[4815]: I0225 13:31:02.044640 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:31:02 crc kubenswrapper[4815]: I0225 13:31:02.081769 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" podStartSLOduration=7.081750884 podStartE2EDuration="7.081750884s" podCreationTimestamp="2026-02-25 13:30:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:31:02.080868875 +0000 UTC m=+619.881966929" watchObservedRunningTime="2026-02-25 13:31:02.081750884 +0000 UTC m=+619.882848928" Feb 25 13:31:02 crc kubenswrapper[4815]: I0225 13:31:02.084944 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:31:03 crc kubenswrapper[4815]: I0225 13:31:03.050617 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:31:03 crc kubenswrapper[4815]: I0225 13:31:03.050652 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:31:03 crc kubenswrapper[4815]: I0225 13:31:03.084177 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:31:04 crc kubenswrapper[4815]: I0225 13:31:04.650070 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-24znl" Feb 25 13:31:04 crc kubenswrapper[4815]: I0225 13:31:04.720207 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-z4lg2"] Feb 25 13:31:21 crc kubenswrapper[4815]: I0225 13:31:21.342580 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:31:21 crc kubenswrapper[4815]: I0225 13:31:21.343349 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:31:21 crc kubenswrapper[4815]: I0225 13:31:21.343408 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:31:21 crc kubenswrapper[4815]: I0225 13:31:21.344147 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e00e443cf7572407e0857913be27d833d029fb2c70b14c349ec691336c7312d"} pod="openshift-machine-config-operator/machine-config-daemon-frc75" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 13:31:21 crc kubenswrapper[4815]: I0225 13:31:21.344206 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" containerID="cri-o://4e00e443cf7572407e0857913be27d833d029fb2c70b14c349ec691336c7312d" gracePeriod=600 Feb 25 13:31:22 crc kubenswrapper[4815]: I0225 13:31:22.187845 4815 generic.go:334] "Generic (PLEG): container finished" podID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerID="4e00e443cf7572407e0857913be27d833d029fb2c70b14c349ec691336c7312d" exitCode=0 Feb 25 13:31:22 crc kubenswrapper[4815]: I0225 13:31:22.188484 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerDied","Data":"4e00e443cf7572407e0857913be27d833d029fb2c70b14c349ec691336c7312d"} Feb 25 13:31:22 crc kubenswrapper[4815]: I0225 13:31:22.188554 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerStarted","Data":"6557b63a0f423077dc6748b76d4eac34119edc35e845268afc85f7e306b0281a"} Feb 25 13:31:22 crc kubenswrapper[4815]: I0225 13:31:22.188582 4815 scope.go:117] "RemoveContainer" containerID="0bde58aaffd292f31f5aeec24700d4368aa27e806f31f168fd069eb408b17058" Feb 25 13:31:25 crc kubenswrapper[4815]: I0225 13:31:25.580386 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qc8ls" Feb 25 13:31:29 crc kubenswrapper[4815]: I0225 13:31:29.768713 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" podUID="072fbe81-88e1-4755-bb59-5ee95d72760f" containerName="registry" containerID="cri-o://4a0b4fa3f4b8e7df475878429dc453ee25d3a6301460b194191752957741cd60" gracePeriod=30 Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.190321 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.250975 4815 generic.go:334] "Generic (PLEG): container finished" podID="072fbe81-88e1-4755-bb59-5ee95d72760f" containerID="4a0b4fa3f4b8e7df475878429dc453ee25d3a6301460b194191752957741cd60" exitCode=0 Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.251030 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" event={"ID":"072fbe81-88e1-4755-bb59-5ee95d72760f","Type":"ContainerDied","Data":"4a0b4fa3f4b8e7df475878429dc453ee25d3a6301460b194191752957741cd60"} Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.251059 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" event={"ID":"072fbe81-88e1-4755-bb59-5ee95d72760f","Type":"ContainerDied","Data":"c7ff748a7a9ef0f49190a50fd5aac69c46afab5e9cc542eee8b6efcf1420a0d6"} Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.251077 4815 scope.go:117] "RemoveContainer" containerID="4a0b4fa3f4b8e7df475878429dc453ee25d3a6301460b194191752957741cd60" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.251198 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-z4lg2" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.262327 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/072fbe81-88e1-4755-bb59-5ee95d72760f-installation-pull-secrets\") pod \"072fbe81-88e1-4755-bb59-5ee95d72760f\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.262664 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"072fbe81-88e1-4755-bb59-5ee95d72760f\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.262738 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/072fbe81-88e1-4755-bb59-5ee95d72760f-trusted-ca\") pod \"072fbe81-88e1-4755-bb59-5ee95d72760f\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.262830 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/072fbe81-88e1-4755-bb59-5ee95d72760f-ca-trust-extracted\") pod \"072fbe81-88e1-4755-bb59-5ee95d72760f\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.262935 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-registry-tls\") pod \"072fbe81-88e1-4755-bb59-5ee95d72760f\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.262971 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/072fbe81-88e1-4755-bb59-5ee95d72760f-registry-certificates\") pod \"072fbe81-88e1-4755-bb59-5ee95d72760f\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.263004 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2l6w\" (UniqueName: \"kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-kube-api-access-x2l6w\") pod \"072fbe81-88e1-4755-bb59-5ee95d72760f\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.263039 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-bound-sa-token\") pod \"072fbe81-88e1-4755-bb59-5ee95d72760f\" (UID: \"072fbe81-88e1-4755-bb59-5ee95d72760f\") " Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.266318 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/072fbe81-88e1-4755-bb59-5ee95d72760f-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "072fbe81-88e1-4755-bb59-5ee95d72760f" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.266944 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/072fbe81-88e1-4755-bb59-5ee95d72760f-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "072fbe81-88e1-4755-bb59-5ee95d72760f" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.269134 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "072fbe81-88e1-4755-bb59-5ee95d72760f" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.270052 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/072fbe81-88e1-4755-bb59-5ee95d72760f-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "072fbe81-88e1-4755-bb59-5ee95d72760f" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.270093 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "072fbe81-88e1-4755-bb59-5ee95d72760f" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.270971 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-kube-api-access-x2l6w" (OuterVolumeSpecName: "kube-api-access-x2l6w") pod "072fbe81-88e1-4755-bb59-5ee95d72760f" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f"). InnerVolumeSpecName "kube-api-access-x2l6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.279448 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "072fbe81-88e1-4755-bb59-5ee95d72760f" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.279584 4815 scope.go:117] "RemoveContainer" containerID="4a0b4fa3f4b8e7df475878429dc453ee25d3a6301460b194191752957741cd60" Feb 25 13:31:30 crc kubenswrapper[4815]: E0225 13:31:30.280198 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a0b4fa3f4b8e7df475878429dc453ee25d3a6301460b194191752957741cd60\": container with ID starting with 4a0b4fa3f4b8e7df475878429dc453ee25d3a6301460b194191752957741cd60 not found: ID does not exist" containerID="4a0b4fa3f4b8e7df475878429dc453ee25d3a6301460b194191752957741cd60" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.280232 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a0b4fa3f4b8e7df475878429dc453ee25d3a6301460b194191752957741cd60"} err="failed to get container status \"4a0b4fa3f4b8e7df475878429dc453ee25d3a6301460b194191752957741cd60\": rpc error: code = NotFound desc = could not find container \"4a0b4fa3f4b8e7df475878429dc453ee25d3a6301460b194191752957741cd60\": container with ID starting with 4a0b4fa3f4b8e7df475878429dc453ee25d3a6301460b194191752957741cd60 not found: ID does not exist" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.298578 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/072fbe81-88e1-4755-bb59-5ee95d72760f-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "072fbe81-88e1-4755-bb59-5ee95d72760f" (UID: "072fbe81-88e1-4755-bb59-5ee95d72760f"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.365426 4815 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/072fbe81-88e1-4755-bb59-5ee95d72760f-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.365480 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/072fbe81-88e1-4755-bb59-5ee95d72760f-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.365499 4815 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/072fbe81-88e1-4755-bb59-5ee95d72760f-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.365541 4815 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.365560 4815 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/072fbe81-88e1-4755-bb59-5ee95d72760f-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.365580 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2l6w\" (UniqueName: \"kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-kube-api-access-x2l6w\") on node \"crc\" DevicePath \"\"" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.365599 4815 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/072fbe81-88e1-4755-bb59-5ee95d72760f-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.624296 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-z4lg2"] Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.627727 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-z4lg2"] Feb 25 13:31:30 crc kubenswrapper[4815]: I0225 13:31:30.948087 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="072fbe81-88e1-4755-bb59-5ee95d72760f" path="/var/lib/kubelet/pods/072fbe81-88e1-4755-bb59-5ee95d72760f/volumes" Feb 25 13:31:32 crc kubenswrapper[4815]: I0225 13:31:32.717588 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp"] Feb 25 13:31:32 crc kubenswrapper[4815]: E0225 13:31:32.718415 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="072fbe81-88e1-4755-bb59-5ee95d72760f" containerName="registry" Feb 25 13:31:32 crc kubenswrapper[4815]: I0225 13:31:32.718448 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="072fbe81-88e1-4755-bb59-5ee95d72760f" containerName="registry" Feb 25 13:31:32 crc kubenswrapper[4815]: I0225 13:31:32.718717 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="072fbe81-88e1-4755-bb59-5ee95d72760f" containerName="registry" Feb 25 13:31:32 crc kubenswrapper[4815]: I0225 13:31:32.720143 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" Feb 25 13:31:32 crc kubenswrapper[4815]: I0225 13:31:32.724271 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 25 13:31:32 crc kubenswrapper[4815]: I0225 13:31:32.725613 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp"] Feb 25 13:31:32 crc kubenswrapper[4815]: I0225 13:31:32.798915 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p572z\" (UniqueName: \"kubernetes.io/projected/6702f60b-0e49-4104-9434-6c3bf4e77cae-kube-api-access-p572z\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp\" (UID: \"6702f60b-0e49-4104-9434-6c3bf4e77cae\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" Feb 25 13:31:32 crc kubenswrapper[4815]: I0225 13:31:32.799001 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6702f60b-0e49-4104-9434-6c3bf4e77cae-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp\" (UID: \"6702f60b-0e49-4104-9434-6c3bf4e77cae\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" Feb 25 13:31:32 crc kubenswrapper[4815]: I0225 13:31:32.799036 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6702f60b-0e49-4104-9434-6c3bf4e77cae-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp\" (UID: \"6702f60b-0e49-4104-9434-6c3bf4e77cae\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" Feb 25 13:31:32 crc kubenswrapper[4815]: I0225 13:31:32.900910 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6702f60b-0e49-4104-9434-6c3bf4e77cae-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp\" (UID: \"6702f60b-0e49-4104-9434-6c3bf4e77cae\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" Feb 25 13:31:32 crc kubenswrapper[4815]: I0225 13:31:32.900980 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6702f60b-0e49-4104-9434-6c3bf4e77cae-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp\" (UID: \"6702f60b-0e49-4104-9434-6c3bf4e77cae\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" Feb 25 13:31:32 crc kubenswrapper[4815]: I0225 13:31:32.901139 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p572z\" (UniqueName: \"kubernetes.io/projected/6702f60b-0e49-4104-9434-6c3bf4e77cae-kube-api-access-p572z\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp\" (UID: \"6702f60b-0e49-4104-9434-6c3bf4e77cae\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" Feb 25 13:31:32 crc kubenswrapper[4815]: I0225 13:31:32.902109 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6702f60b-0e49-4104-9434-6c3bf4e77cae-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp\" (UID: \"6702f60b-0e49-4104-9434-6c3bf4e77cae\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" Feb 25 13:31:32 crc kubenswrapper[4815]: I0225 13:31:32.902195 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6702f60b-0e49-4104-9434-6c3bf4e77cae-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp\" (UID: \"6702f60b-0e49-4104-9434-6c3bf4e77cae\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" Feb 25 13:31:32 crc kubenswrapper[4815]: I0225 13:31:32.934825 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p572z\" (UniqueName: \"kubernetes.io/projected/6702f60b-0e49-4104-9434-6c3bf4e77cae-kube-api-access-p572z\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp\" (UID: \"6702f60b-0e49-4104-9434-6c3bf4e77cae\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" Feb 25 13:31:33 crc kubenswrapper[4815]: I0225 13:31:33.074281 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" Feb 25 13:31:33 crc kubenswrapper[4815]: I0225 13:31:33.515780 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp"] Feb 25 13:31:34 crc kubenswrapper[4815]: I0225 13:31:34.285456 4815 generic.go:334] "Generic (PLEG): container finished" podID="6702f60b-0e49-4104-9434-6c3bf4e77cae" containerID="46ae3dd45b250bdcfefa577ca04acba06629ec21f7cb1aec58d85b768046cb7b" exitCode=0 Feb 25 13:31:34 crc kubenswrapper[4815]: I0225 13:31:34.285570 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" event={"ID":"6702f60b-0e49-4104-9434-6c3bf4e77cae","Type":"ContainerDied","Data":"46ae3dd45b250bdcfefa577ca04acba06629ec21f7cb1aec58d85b768046cb7b"} Feb 25 13:31:34 crc kubenswrapper[4815]: I0225 13:31:34.285644 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" event={"ID":"6702f60b-0e49-4104-9434-6c3bf4e77cae","Type":"ContainerStarted","Data":"5d3f7c6d93f44c76c1dcf459ce49a400ac3d429f6a26637589b3f06c7a0d3d4b"} Feb 25 13:31:36 crc kubenswrapper[4815]: I0225 13:31:36.298692 4815 generic.go:334] "Generic (PLEG): container finished" podID="6702f60b-0e49-4104-9434-6c3bf4e77cae" containerID="049af07f611ea7f5ad1fe7542cb399fe5810b3c9e0894115a2bf71ed0d08e82e" exitCode=0 Feb 25 13:31:36 crc kubenswrapper[4815]: I0225 13:31:36.298807 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" event={"ID":"6702f60b-0e49-4104-9434-6c3bf4e77cae","Type":"ContainerDied","Data":"049af07f611ea7f5ad1fe7542cb399fe5810b3c9e0894115a2bf71ed0d08e82e"} Feb 25 13:31:37 crc kubenswrapper[4815]: I0225 13:31:37.307138 4815 generic.go:334] "Generic (PLEG): container finished" podID="6702f60b-0e49-4104-9434-6c3bf4e77cae" containerID="1b32ca1796897fe8fcabd35d6f1052cbbc501ff012a785c3c37e977e64f473d2" exitCode=0 Feb 25 13:31:37 crc kubenswrapper[4815]: I0225 13:31:37.307198 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" event={"ID":"6702f60b-0e49-4104-9434-6c3bf4e77cae","Type":"ContainerDied","Data":"1b32ca1796897fe8fcabd35d6f1052cbbc501ff012a785c3c37e977e64f473d2"} Feb 25 13:31:38 crc kubenswrapper[4815]: I0225 13:31:38.562885 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" Feb 25 13:31:38 crc kubenswrapper[4815]: I0225 13:31:38.681965 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6702f60b-0e49-4104-9434-6c3bf4e77cae-bundle\") pod \"6702f60b-0e49-4104-9434-6c3bf4e77cae\" (UID: \"6702f60b-0e49-4104-9434-6c3bf4e77cae\") " Feb 25 13:31:38 crc kubenswrapper[4815]: I0225 13:31:38.682478 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6702f60b-0e49-4104-9434-6c3bf4e77cae-util\") pod \"6702f60b-0e49-4104-9434-6c3bf4e77cae\" (UID: \"6702f60b-0e49-4104-9434-6c3bf4e77cae\") " Feb 25 13:31:38 crc kubenswrapper[4815]: I0225 13:31:38.682630 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p572z\" (UniqueName: \"kubernetes.io/projected/6702f60b-0e49-4104-9434-6c3bf4e77cae-kube-api-access-p572z\") pod \"6702f60b-0e49-4104-9434-6c3bf4e77cae\" (UID: \"6702f60b-0e49-4104-9434-6c3bf4e77cae\") " Feb 25 13:31:38 crc kubenswrapper[4815]: I0225 13:31:38.682851 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6702f60b-0e49-4104-9434-6c3bf4e77cae-bundle" (OuterVolumeSpecName: "bundle") pod "6702f60b-0e49-4104-9434-6c3bf4e77cae" (UID: "6702f60b-0e49-4104-9434-6c3bf4e77cae"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:31:38 crc kubenswrapper[4815]: I0225 13:31:38.683116 4815 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6702f60b-0e49-4104-9434-6c3bf4e77cae-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:31:38 crc kubenswrapper[4815]: I0225 13:31:38.687823 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6702f60b-0e49-4104-9434-6c3bf4e77cae-kube-api-access-p572z" (OuterVolumeSpecName: "kube-api-access-p572z") pod "6702f60b-0e49-4104-9434-6c3bf4e77cae" (UID: "6702f60b-0e49-4104-9434-6c3bf4e77cae"). InnerVolumeSpecName "kube-api-access-p572z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:31:38 crc kubenswrapper[4815]: I0225 13:31:38.696459 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6702f60b-0e49-4104-9434-6c3bf4e77cae-util" (OuterVolumeSpecName: "util") pod "6702f60b-0e49-4104-9434-6c3bf4e77cae" (UID: "6702f60b-0e49-4104-9434-6c3bf4e77cae"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:31:38 crc kubenswrapper[4815]: I0225 13:31:38.786200 4815 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6702f60b-0e49-4104-9434-6c3bf4e77cae-util\") on node \"crc\" DevicePath \"\"" Feb 25 13:31:38 crc kubenswrapper[4815]: I0225 13:31:38.786268 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p572z\" (UniqueName: \"kubernetes.io/projected/6702f60b-0e49-4104-9434-6c3bf4e77cae-kube-api-access-p572z\") on node \"crc\" DevicePath \"\"" Feb 25 13:31:39 crc kubenswrapper[4815]: I0225 13:31:39.333414 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" event={"ID":"6702f60b-0e49-4104-9434-6c3bf4e77cae","Type":"ContainerDied","Data":"5d3f7c6d93f44c76c1dcf459ce49a400ac3d429f6a26637589b3f06c7a0d3d4b"} Feb 25 13:31:39 crc kubenswrapper[4815]: I0225 13:31:39.333499 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d3f7c6d93f44c76c1dcf459ce49a400ac3d429f6a26637589b3f06c7a0d3d4b" Feb 25 13:31:39 crc kubenswrapper[4815]: I0225 13:31:39.333693 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp" Feb 25 13:31:41 crc kubenswrapper[4815]: I0225 13:31:41.410024 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-xwh42"] Feb 25 13:31:41 crc kubenswrapper[4815]: E0225 13:31:41.410230 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6702f60b-0e49-4104-9434-6c3bf4e77cae" containerName="pull" Feb 25 13:31:41 crc kubenswrapper[4815]: I0225 13:31:41.410241 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6702f60b-0e49-4104-9434-6c3bf4e77cae" containerName="pull" Feb 25 13:31:41 crc kubenswrapper[4815]: E0225 13:31:41.410254 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6702f60b-0e49-4104-9434-6c3bf4e77cae" containerName="util" Feb 25 13:31:41 crc kubenswrapper[4815]: I0225 13:31:41.410260 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6702f60b-0e49-4104-9434-6c3bf4e77cae" containerName="util" Feb 25 13:31:41 crc kubenswrapper[4815]: E0225 13:31:41.410267 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6702f60b-0e49-4104-9434-6c3bf4e77cae" containerName="extract" Feb 25 13:31:41 crc kubenswrapper[4815]: I0225 13:31:41.410273 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6702f60b-0e49-4104-9434-6c3bf4e77cae" containerName="extract" Feb 25 13:31:41 crc kubenswrapper[4815]: I0225 13:31:41.410359 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="6702f60b-0e49-4104-9434-6c3bf4e77cae" containerName="extract" Feb 25 13:31:41 crc kubenswrapper[4815]: I0225 13:31:41.410744 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-xwh42" Feb 25 13:31:41 crc kubenswrapper[4815]: I0225 13:31:41.413245 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 25 13:31:41 crc kubenswrapper[4815]: I0225 13:31:41.413412 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 25 13:31:41 crc kubenswrapper[4815]: I0225 13:31:41.413915 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-mhstd" Feb 25 13:31:41 crc kubenswrapper[4815]: I0225 13:31:41.427529 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-xwh42"] Feb 25 13:31:41 crc kubenswrapper[4815]: I0225 13:31:41.519241 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjgkt\" (UniqueName: \"kubernetes.io/projected/ffbc3f3b-118b-4b77-b3a1-b8b20201a08b-kube-api-access-fjgkt\") pod \"nmstate-operator-694c9596b7-xwh42\" (UID: \"ffbc3f3b-118b-4b77-b3a1-b8b20201a08b\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-xwh42" Feb 25 13:31:41 crc kubenswrapper[4815]: I0225 13:31:41.619970 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjgkt\" (UniqueName: \"kubernetes.io/projected/ffbc3f3b-118b-4b77-b3a1-b8b20201a08b-kube-api-access-fjgkt\") pod \"nmstate-operator-694c9596b7-xwh42\" (UID: \"ffbc3f3b-118b-4b77-b3a1-b8b20201a08b\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-xwh42" Feb 25 13:31:41 crc kubenswrapper[4815]: I0225 13:31:41.639087 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjgkt\" (UniqueName: \"kubernetes.io/projected/ffbc3f3b-118b-4b77-b3a1-b8b20201a08b-kube-api-access-fjgkt\") pod \"nmstate-operator-694c9596b7-xwh42\" (UID: \"ffbc3f3b-118b-4b77-b3a1-b8b20201a08b\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-xwh42" Feb 25 13:31:41 crc kubenswrapper[4815]: I0225 13:31:41.726370 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-xwh42" Feb 25 13:31:42 crc kubenswrapper[4815]: I0225 13:31:42.147065 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-xwh42"] Feb 25 13:31:42 crc kubenswrapper[4815]: I0225 13:31:42.347399 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-xwh42" event={"ID":"ffbc3f3b-118b-4b77-b3a1-b8b20201a08b","Type":"ContainerStarted","Data":"8e440797c909e285e62b668821a9f952be7ccb68d244ecaa6b287e6c179c44e8"} Feb 25 13:31:46 crc kubenswrapper[4815]: I0225 13:31:46.370978 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-xwh42" event={"ID":"ffbc3f3b-118b-4b77-b3a1-b8b20201a08b","Type":"ContainerStarted","Data":"461d01f8b681c7597060eed0c2ffb29edf11e243d0e5e76473f005df2474f8f3"} Feb 25 13:31:46 crc kubenswrapper[4815]: I0225 13:31:46.387307 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-694c9596b7-xwh42" podStartSLOduration=2.136618516 podStartE2EDuration="5.387285496s" podCreationTimestamp="2026-02-25 13:31:41 +0000 UTC" firstStartedPulling="2026-02-25 13:31:42.155240825 +0000 UTC m=+659.956338879" lastFinishedPulling="2026-02-25 13:31:45.405907805 +0000 UTC m=+663.207005859" observedRunningTime="2026-02-25 13:31:46.384541188 +0000 UTC m=+664.185639282" watchObservedRunningTime="2026-02-25 13:31:46.387285496 +0000 UTC m=+664.188383550" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.232992 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-ljnd9"] Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.234286 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-ljnd9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.237333 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-rgx6w" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.246955 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-ljnd9"] Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.280472 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-p9cpt"] Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.281301 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-p9cpt" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.288282 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-sgk4q"] Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.288468 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brhm8\" (UniqueName: \"kubernetes.io/projected/40f20bd3-e16c-490e-9316-fa6c6e36ab84-kube-api-access-brhm8\") pod \"nmstate-metrics-58c85c668d-ljnd9\" (UID: \"40f20bd3-e16c-490e-9316-fa6c6e36ab84\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-ljnd9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.294987 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-sgk4q" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.297487 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.311284 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-sgk4q"] Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.383839 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg"] Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.384576 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.386273 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-bh7p7" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.386711 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.387125 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.389421 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg"] Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.393946 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/43c7d725-8292-4e17-bc55-bfb18a7bc51e-ovs-socket\") pod \"nmstate-handler-p9cpt\" (UID: \"43c7d725-8292-4e17-bc55-bfb18a7bc51e\") " pod="openshift-nmstate/nmstate-handler-p9cpt" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.394005 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/43c7d725-8292-4e17-bc55-bfb18a7bc51e-nmstate-lock\") pod \"nmstate-handler-p9cpt\" (UID: \"43c7d725-8292-4e17-bc55-bfb18a7bc51e\") " pod="openshift-nmstate/nmstate-handler-p9cpt" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.394055 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/266d6bc5-8e54-4af2-b6e4-5b3884ab882e-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-sgk4q\" (UID: \"266d6bc5-8e54-4af2-b6e4-5b3884ab882e\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-sgk4q" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.394092 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/43c7d725-8292-4e17-bc55-bfb18a7bc51e-dbus-socket\") pod \"nmstate-handler-p9cpt\" (UID: \"43c7d725-8292-4e17-bc55-bfb18a7bc51e\") " pod="openshift-nmstate/nmstate-handler-p9cpt" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.394117 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzpkk\" (UniqueName: \"kubernetes.io/projected/266d6bc5-8e54-4af2-b6e4-5b3884ab882e-kube-api-access-zzpkk\") pod \"nmstate-webhook-866bcb46dc-sgk4q\" (UID: \"266d6bc5-8e54-4af2-b6e4-5b3884ab882e\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-sgk4q" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.394148 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xks6b\" (UniqueName: \"kubernetes.io/projected/43c7d725-8292-4e17-bc55-bfb18a7bc51e-kube-api-access-xks6b\") pod \"nmstate-handler-p9cpt\" (UID: \"43c7d725-8292-4e17-bc55-bfb18a7bc51e\") " pod="openshift-nmstate/nmstate-handler-p9cpt" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.394177 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brhm8\" (UniqueName: \"kubernetes.io/projected/40f20bd3-e16c-490e-9316-fa6c6e36ab84-kube-api-access-brhm8\") pod \"nmstate-metrics-58c85c668d-ljnd9\" (UID: \"40f20bd3-e16c-490e-9316-fa6c6e36ab84\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-ljnd9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.417145 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brhm8\" (UniqueName: \"kubernetes.io/projected/40f20bd3-e16c-490e-9316-fa6c6e36ab84-kube-api-access-brhm8\") pod \"nmstate-metrics-58c85c668d-ljnd9\" (UID: \"40f20bd3-e16c-490e-9316-fa6c6e36ab84\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-ljnd9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.495794 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/266d6bc5-8e54-4af2-b6e4-5b3884ab882e-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-sgk4q\" (UID: \"266d6bc5-8e54-4af2-b6e4-5b3884ab882e\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-sgk4q" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.495865 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/bb89c9f3-b457-466d-a6a5-799ab144b125-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-wbxlg\" (UID: \"bb89c9f3-b457-466d-a6a5-799ab144b125\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.495893 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/43c7d725-8292-4e17-bc55-bfb18a7bc51e-dbus-socket\") pod \"nmstate-handler-p9cpt\" (UID: \"43c7d725-8292-4e17-bc55-bfb18a7bc51e\") " pod="openshift-nmstate/nmstate-handler-p9cpt" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.495912 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzpkk\" (UniqueName: \"kubernetes.io/projected/266d6bc5-8e54-4af2-b6e4-5b3884ab882e-kube-api-access-zzpkk\") pod \"nmstate-webhook-866bcb46dc-sgk4q\" (UID: \"266d6bc5-8e54-4af2-b6e4-5b3884ab882e\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-sgk4q" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.495934 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xks6b\" (UniqueName: \"kubernetes.io/projected/43c7d725-8292-4e17-bc55-bfb18a7bc51e-kube-api-access-xks6b\") pod \"nmstate-handler-p9cpt\" (UID: \"43c7d725-8292-4e17-bc55-bfb18a7bc51e\") " pod="openshift-nmstate/nmstate-handler-p9cpt" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.495964 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/43c7d725-8292-4e17-bc55-bfb18a7bc51e-ovs-socket\") pod \"nmstate-handler-p9cpt\" (UID: \"43c7d725-8292-4e17-bc55-bfb18a7bc51e\") " pod="openshift-nmstate/nmstate-handler-p9cpt" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.495984 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcldk\" (UniqueName: \"kubernetes.io/projected/bb89c9f3-b457-466d-a6a5-799ab144b125-kube-api-access-lcldk\") pod \"nmstate-console-plugin-5c78fc5d65-wbxlg\" (UID: \"bb89c9f3-b457-466d-a6a5-799ab144b125\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.496018 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb89c9f3-b457-466d-a6a5-799ab144b125-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-wbxlg\" (UID: \"bb89c9f3-b457-466d-a6a5-799ab144b125\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.496036 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/43c7d725-8292-4e17-bc55-bfb18a7bc51e-nmstate-lock\") pod \"nmstate-handler-p9cpt\" (UID: \"43c7d725-8292-4e17-bc55-bfb18a7bc51e\") " pod="openshift-nmstate/nmstate-handler-p9cpt" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.496114 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/43c7d725-8292-4e17-bc55-bfb18a7bc51e-nmstate-lock\") pod \"nmstate-handler-p9cpt\" (UID: \"43c7d725-8292-4e17-bc55-bfb18a7bc51e\") " pod="openshift-nmstate/nmstate-handler-p9cpt" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.496279 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/43c7d725-8292-4e17-bc55-bfb18a7bc51e-dbus-socket\") pod \"nmstate-handler-p9cpt\" (UID: \"43c7d725-8292-4e17-bc55-bfb18a7bc51e\") " pod="openshift-nmstate/nmstate-handler-p9cpt" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.496490 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/43c7d725-8292-4e17-bc55-bfb18a7bc51e-ovs-socket\") pod \"nmstate-handler-p9cpt\" (UID: \"43c7d725-8292-4e17-bc55-bfb18a7bc51e\") " pod="openshift-nmstate/nmstate-handler-p9cpt" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.500593 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/266d6bc5-8e54-4af2-b6e4-5b3884ab882e-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-sgk4q\" (UID: \"266d6bc5-8e54-4af2-b6e4-5b3884ab882e\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-sgk4q" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.514224 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xks6b\" (UniqueName: \"kubernetes.io/projected/43c7d725-8292-4e17-bc55-bfb18a7bc51e-kube-api-access-xks6b\") pod \"nmstate-handler-p9cpt\" (UID: \"43c7d725-8292-4e17-bc55-bfb18a7bc51e\") " pod="openshift-nmstate/nmstate-handler-p9cpt" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.520114 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzpkk\" (UniqueName: \"kubernetes.io/projected/266d6bc5-8e54-4af2-b6e4-5b3884ab882e-kube-api-access-zzpkk\") pod \"nmstate-webhook-866bcb46dc-sgk4q\" (UID: \"266d6bc5-8e54-4af2-b6e4-5b3884ab882e\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-sgk4q" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.564384 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-ljnd9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.569464 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5b8fdc85f8-zvtg9"] Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.570116 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.588800 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5b8fdc85f8-zvtg9"] Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.597612 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb89c9f3-b457-466d-a6a5-799ab144b125-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-wbxlg\" (UID: \"bb89c9f3-b457-466d-a6a5-799ab144b125\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.597683 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/bb89c9f3-b457-466d-a6a5-799ab144b125-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-wbxlg\" (UID: \"bb89c9f3-b457-466d-a6a5-799ab144b125\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.597754 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcldk\" (UniqueName: \"kubernetes.io/projected/bb89c9f3-b457-466d-a6a5-799ab144b125-kube-api-access-lcldk\") pod \"nmstate-console-plugin-5c78fc5d65-wbxlg\" (UID: \"bb89c9f3-b457-466d-a6a5-799ab144b125\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg" Feb 25 13:31:47 crc kubenswrapper[4815]: E0225 13:31:47.598076 4815 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Feb 25 13:31:47 crc kubenswrapper[4815]: E0225 13:31:47.598122 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb89c9f3-b457-466d-a6a5-799ab144b125-plugin-serving-cert podName:bb89c9f3-b457-466d-a6a5-799ab144b125 nodeName:}" failed. No retries permitted until 2026-02-25 13:31:48.098107707 +0000 UTC m=+665.899205761 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/bb89c9f3-b457-466d-a6a5-799ab144b125-plugin-serving-cert") pod "nmstate-console-plugin-5c78fc5d65-wbxlg" (UID: "bb89c9f3-b457-466d-a6a5-799ab144b125") : secret "plugin-serving-cert" not found Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.599272 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/bb89c9f3-b457-466d-a6a5-799ab144b125-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-wbxlg\" (UID: \"bb89c9f3-b457-466d-a6a5-799ab144b125\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.599471 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-p9cpt" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.617072 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-sgk4q" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.623355 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcldk\" (UniqueName: \"kubernetes.io/projected/bb89c9f3-b457-466d-a6a5-799ab144b125-kube-api-access-lcldk\") pod \"nmstate-console-plugin-5c78fc5d65-wbxlg\" (UID: \"bb89c9f3-b457-466d-a6a5-799ab144b125\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.698740 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-service-ca\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.698781 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxdkp\" (UniqueName: \"kubernetes.io/projected/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-kube-api-access-jxdkp\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.698827 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-console-oauth-config\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.698850 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-console-config\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.698879 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-oauth-serving-cert\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.698910 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-trusted-ca-bundle\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.698944 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-console-serving-cert\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.800533 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-console-config\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.800592 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-oauth-serving-cert\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.800626 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-trusted-ca-bundle\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.800659 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-console-serving-cert\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.800692 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-service-ca\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.800710 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxdkp\" (UniqueName: \"kubernetes.io/projected/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-kube-api-access-jxdkp\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.800744 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-console-oauth-config\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.801705 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-console-config\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.802102 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-trusted-ca-bundle\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.802345 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-service-ca\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.802378 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-oauth-serving-cert\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.808724 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-console-serving-cert\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.810592 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-console-oauth-config\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.817596 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxdkp\" (UniqueName: \"kubernetes.io/projected/cbaaed81-4820-45aa-bf7e-9230e15cc5b3-kube-api-access-jxdkp\") pod \"console-5b8fdc85f8-zvtg9\" (UID: \"cbaaed81-4820-45aa-bf7e-9230e15cc5b3\") " pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.844991 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-sgk4q"] Feb 25 13:31:47 crc kubenswrapper[4815]: I0225 13:31:47.917488 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:48 crc kubenswrapper[4815]: I0225 13:31:48.103876 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb89c9f3-b457-466d-a6a5-799ab144b125-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-wbxlg\" (UID: \"bb89c9f3-b457-466d-a6a5-799ab144b125\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg" Feb 25 13:31:48 crc kubenswrapper[4815]: I0225 13:31:48.107667 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb89c9f3-b457-466d-a6a5-799ab144b125-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-wbxlg\" (UID: \"bb89c9f3-b457-466d-a6a5-799ab144b125\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg" Feb 25 13:31:48 crc kubenswrapper[4815]: I0225 13:31:48.134106 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-ljnd9"] Feb 25 13:31:48 crc kubenswrapper[4815]: I0225 13:31:48.311776 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg" Feb 25 13:31:48 crc kubenswrapper[4815]: I0225 13:31:48.396360 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-sgk4q" event={"ID":"266d6bc5-8e54-4af2-b6e4-5b3884ab882e","Type":"ContainerStarted","Data":"7f5c7652ca51d026c07000a0bda84fb6597b6f08cee4b0aebcfcfe5299d4f78b"} Feb 25 13:31:48 crc kubenswrapper[4815]: I0225 13:31:48.398000 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-ljnd9" event={"ID":"40f20bd3-e16c-490e-9316-fa6c6e36ab84","Type":"ContainerStarted","Data":"5383d6b16154e25783b2177c89878a907a2f760c5c618e48621d59bbb65884cf"} Feb 25 13:31:48 crc kubenswrapper[4815]: I0225 13:31:48.399678 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-p9cpt" event={"ID":"43c7d725-8292-4e17-bc55-bfb18a7bc51e","Type":"ContainerStarted","Data":"5192be15a3ef04cb0eeccfba5d1fcb4f332bacf1f51d864f898976115ef27120"} Feb 25 13:31:48 crc kubenswrapper[4815]: I0225 13:31:48.415715 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5b8fdc85f8-zvtg9"] Feb 25 13:31:48 crc kubenswrapper[4815]: W0225 13:31:48.445563 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbaaed81_4820_45aa_bf7e_9230e15cc5b3.slice/crio-76bdf54954e7b20490f0b489331deedb71aa824380d825a86e8e476dcfb4d712 WatchSource:0}: Error finding container 76bdf54954e7b20490f0b489331deedb71aa824380d825a86e8e476dcfb4d712: Status 404 returned error can't find the container with id 76bdf54954e7b20490f0b489331deedb71aa824380d825a86e8e476dcfb4d712 Feb 25 13:31:48 crc kubenswrapper[4815]: I0225 13:31:48.479100 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg"] Feb 25 13:31:48 crc kubenswrapper[4815]: W0225 13:31:48.484931 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb89c9f3_b457_466d_a6a5_799ab144b125.slice/crio-129769af6a4bb248b13c03ce4f7455db44c69869b8305954c736acd6885ec414 WatchSource:0}: Error finding container 129769af6a4bb248b13c03ce4f7455db44c69869b8305954c736acd6885ec414: Status 404 returned error can't find the container with id 129769af6a4bb248b13c03ce4f7455db44c69869b8305954c736acd6885ec414 Feb 25 13:31:49 crc kubenswrapper[4815]: I0225 13:31:49.405655 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg" event={"ID":"bb89c9f3-b457-466d-a6a5-799ab144b125","Type":"ContainerStarted","Data":"129769af6a4bb248b13c03ce4f7455db44c69869b8305954c736acd6885ec414"} Feb 25 13:31:49 crc kubenswrapper[4815]: I0225 13:31:49.407527 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5b8fdc85f8-zvtg9" event={"ID":"cbaaed81-4820-45aa-bf7e-9230e15cc5b3","Type":"ContainerStarted","Data":"76bdf54954e7b20490f0b489331deedb71aa824380d825a86e8e476dcfb4d712"} Feb 25 13:31:50 crc kubenswrapper[4815]: I0225 13:31:50.413119 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5b8fdc85f8-zvtg9" event={"ID":"cbaaed81-4820-45aa-bf7e-9230e15cc5b3","Type":"ContainerStarted","Data":"281c10429cd1261ac042dc5bc7221c60a174cbdae258e4afbbde3392a2f557ed"} Feb 25 13:31:50 crc kubenswrapper[4815]: I0225 13:31:50.434310 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5b8fdc85f8-zvtg9" podStartSLOduration=3.434292276 podStartE2EDuration="3.434292276s" podCreationTimestamp="2026-02-25 13:31:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:31:50.430118773 +0000 UTC m=+668.231216827" watchObservedRunningTime="2026-02-25 13:31:50.434292276 +0000 UTC m=+668.235390330" Feb 25 13:31:52 crc kubenswrapper[4815]: I0225 13:31:52.424395 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-ljnd9" event={"ID":"40f20bd3-e16c-490e-9316-fa6c6e36ab84","Type":"ContainerStarted","Data":"ae75dae9003aa8f4e46ddc77eb86fe5177b2a87ef37559a665cfb3b5b65690e0"} Feb 25 13:31:52 crc kubenswrapper[4815]: I0225 13:31:52.426340 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg" event={"ID":"bb89c9f3-b457-466d-a6a5-799ab144b125","Type":"ContainerStarted","Data":"1748d8f0dfcc01d3764ac9749eb518ba8cdbc1dfc640a056467f63a2001cadcd"} Feb 25 13:31:52 crc kubenswrapper[4815]: I0225 13:31:52.427226 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-p9cpt" event={"ID":"43c7d725-8292-4e17-bc55-bfb18a7bc51e","Type":"ContainerStarted","Data":"c8365d53f87605cbbdcf394c51fa7e1a2684e4da8121752c6315bfbe983b093e"} Feb 25 13:31:52 crc kubenswrapper[4815]: I0225 13:31:52.427356 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-p9cpt" Feb 25 13:31:52 crc kubenswrapper[4815]: I0225 13:31:52.428451 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-sgk4q" event={"ID":"266d6bc5-8e54-4af2-b6e4-5b3884ab882e","Type":"ContainerStarted","Data":"7c8c2d4f590bd6f80133c6a3385b413397d0f5400df9f56f0a8360de94e01013"} Feb 25 13:31:52 crc kubenswrapper[4815]: I0225 13:31:52.428555 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-sgk4q" Feb 25 13:31:52 crc kubenswrapper[4815]: I0225 13:31:52.444943 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-wbxlg" podStartSLOduration=1.962868085 podStartE2EDuration="5.444925389s" podCreationTimestamp="2026-02-25 13:31:47 +0000 UTC" firstStartedPulling="2026-02-25 13:31:48.493067474 +0000 UTC m=+666.294165528" lastFinishedPulling="2026-02-25 13:31:51.975124768 +0000 UTC m=+669.776222832" observedRunningTime="2026-02-25 13:31:52.442196162 +0000 UTC m=+670.243294206" watchObservedRunningTime="2026-02-25 13:31:52.444925389 +0000 UTC m=+670.246023453" Feb 25 13:31:52 crc kubenswrapper[4815]: I0225 13:31:52.461128 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-p9cpt" podStartSLOduration=1.044617514 podStartE2EDuration="5.461107027s" podCreationTimestamp="2026-02-25 13:31:47 +0000 UTC" firstStartedPulling="2026-02-25 13:31:47.628582463 +0000 UTC m=+665.429680507" lastFinishedPulling="2026-02-25 13:31:52.045071966 +0000 UTC m=+669.846170020" observedRunningTime="2026-02-25 13:31:52.459078152 +0000 UTC m=+670.260176206" watchObservedRunningTime="2026-02-25 13:31:52.461107027 +0000 UTC m=+670.262205111" Feb 25 13:31:52 crc kubenswrapper[4815]: I0225 13:31:52.473552 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-sgk4q" podStartSLOduration=1.245947757 podStartE2EDuration="5.473537255s" podCreationTimestamp="2026-02-25 13:31:47 +0000 UTC" firstStartedPulling="2026-02-25 13:31:47.851193306 +0000 UTC m=+665.652291360" lastFinishedPulling="2026-02-25 13:31:52.078782804 +0000 UTC m=+669.879880858" observedRunningTime="2026-02-25 13:31:52.470545609 +0000 UTC m=+670.271643653" watchObservedRunningTime="2026-02-25 13:31:52.473537255 +0000 UTC m=+670.274635309" Feb 25 13:31:56 crc kubenswrapper[4815]: I0225 13:31:56.454480 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-ljnd9" event={"ID":"40f20bd3-e16c-490e-9316-fa6c6e36ab84","Type":"ContainerStarted","Data":"b6628338aaffe4c893cadb7d5914afc606095d805786a26641eb69eab2cdfa65"} Feb 25 13:31:56 crc kubenswrapper[4815]: I0225 13:31:56.479258 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58c85c668d-ljnd9" podStartSLOduration=2.179643302 podStartE2EDuration="9.479224463s" podCreationTimestamp="2026-02-25 13:31:47 +0000 UTC" firstStartedPulling="2026-02-25 13:31:48.134650916 +0000 UTC m=+665.935748970" lastFinishedPulling="2026-02-25 13:31:55.434232077 +0000 UTC m=+673.235330131" observedRunningTime="2026-02-25 13:31:56.474072358 +0000 UTC m=+674.275170492" watchObservedRunningTime="2026-02-25 13:31:56.479224463 +0000 UTC m=+674.280322567" Feb 25 13:31:57 crc kubenswrapper[4815]: I0225 13:31:57.619829 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-p9cpt" Feb 25 13:31:57 crc kubenswrapper[4815]: I0225 13:31:57.918675 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:57 crc kubenswrapper[4815]: I0225 13:31:57.918732 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:57 crc kubenswrapper[4815]: I0225 13:31:57.923844 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:58 crc kubenswrapper[4815]: I0225 13:31:58.473873 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5b8fdc85f8-zvtg9" Feb 25 13:31:58 crc kubenswrapper[4815]: I0225 13:31:58.533592 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-mhgxd"] Feb 25 13:32:00 crc kubenswrapper[4815]: I0225 13:32:00.132163 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533772-htrc4"] Feb 25 13:32:00 crc kubenswrapper[4815]: I0225 13:32:00.133019 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533772-htrc4" Feb 25 13:32:00 crc kubenswrapper[4815]: I0225 13:32:00.137474 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:32:00 crc kubenswrapper[4815]: I0225 13:32:00.137781 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:32:00 crc kubenswrapper[4815]: I0225 13:32:00.138164 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:32:00 crc kubenswrapper[4815]: I0225 13:32:00.145913 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533772-htrc4"] Feb 25 13:32:00 crc kubenswrapper[4815]: I0225 13:32:00.275175 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4x8g\" (UniqueName: \"kubernetes.io/projected/a052dc18-c68b-4e41-8666-0d40b515d99b-kube-api-access-h4x8g\") pod \"auto-csr-approver-29533772-htrc4\" (UID: \"a052dc18-c68b-4e41-8666-0d40b515d99b\") " pod="openshift-infra/auto-csr-approver-29533772-htrc4" Feb 25 13:32:00 crc kubenswrapper[4815]: I0225 13:32:00.377012 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4x8g\" (UniqueName: \"kubernetes.io/projected/a052dc18-c68b-4e41-8666-0d40b515d99b-kube-api-access-h4x8g\") pod \"auto-csr-approver-29533772-htrc4\" (UID: \"a052dc18-c68b-4e41-8666-0d40b515d99b\") " pod="openshift-infra/auto-csr-approver-29533772-htrc4" Feb 25 13:32:00 crc kubenswrapper[4815]: I0225 13:32:00.412273 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4x8g\" (UniqueName: \"kubernetes.io/projected/a052dc18-c68b-4e41-8666-0d40b515d99b-kube-api-access-h4x8g\") pod \"auto-csr-approver-29533772-htrc4\" (UID: \"a052dc18-c68b-4e41-8666-0d40b515d99b\") " pod="openshift-infra/auto-csr-approver-29533772-htrc4" Feb 25 13:32:00 crc kubenswrapper[4815]: I0225 13:32:00.462524 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533772-htrc4" Feb 25 13:32:00 crc kubenswrapper[4815]: I0225 13:32:00.863680 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533772-htrc4"] Feb 25 13:32:00 crc kubenswrapper[4815]: W0225 13:32:00.869748 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda052dc18_c68b_4e41_8666_0d40b515d99b.slice/crio-faa3a10092648070370243fec3b647fe7143b836c0966d5e606b160de289b433 WatchSource:0}: Error finding container faa3a10092648070370243fec3b647fe7143b836c0966d5e606b160de289b433: Status 404 returned error can't find the container with id faa3a10092648070370243fec3b647fe7143b836c0966d5e606b160de289b433 Feb 25 13:32:01 crc kubenswrapper[4815]: I0225 13:32:01.484677 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533772-htrc4" event={"ID":"a052dc18-c68b-4e41-8666-0d40b515d99b","Type":"ContainerStarted","Data":"faa3a10092648070370243fec3b647fe7143b836c0966d5e606b160de289b433"} Feb 25 13:32:02 crc kubenswrapper[4815]: I0225 13:32:02.490438 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533772-htrc4" event={"ID":"a052dc18-c68b-4e41-8666-0d40b515d99b","Type":"ContainerStarted","Data":"54171993046dd1f6abdbe71ffe2e0962df40a595a20f014f4d7593c4403cfcb9"} Feb 25 13:32:02 crc kubenswrapper[4815]: I0225 13:32:02.512075 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533772-htrc4" podStartSLOduration=1.192839162 podStartE2EDuration="2.512054403s" podCreationTimestamp="2026-02-25 13:32:00 +0000 UTC" firstStartedPulling="2026-02-25 13:32:00.872034578 +0000 UTC m=+678.673132672" lastFinishedPulling="2026-02-25 13:32:02.191249819 +0000 UTC m=+679.992347913" observedRunningTime="2026-02-25 13:32:02.506182415 +0000 UTC m=+680.307280469" watchObservedRunningTime="2026-02-25 13:32:02.512054403 +0000 UTC m=+680.313152467" Feb 25 13:32:03 crc kubenswrapper[4815]: I0225 13:32:03.498970 4815 generic.go:334] "Generic (PLEG): container finished" podID="a052dc18-c68b-4e41-8666-0d40b515d99b" containerID="54171993046dd1f6abdbe71ffe2e0962df40a595a20f014f4d7593c4403cfcb9" exitCode=0 Feb 25 13:32:03 crc kubenswrapper[4815]: I0225 13:32:03.499032 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533772-htrc4" event={"ID":"a052dc18-c68b-4e41-8666-0d40b515d99b","Type":"ContainerDied","Data":"54171993046dd1f6abdbe71ffe2e0962df40a595a20f014f4d7593c4403cfcb9"} Feb 25 13:32:04 crc kubenswrapper[4815]: I0225 13:32:04.810267 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533772-htrc4" Feb 25 13:32:04 crc kubenswrapper[4815]: I0225 13:32:04.943013 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4x8g\" (UniqueName: \"kubernetes.io/projected/a052dc18-c68b-4e41-8666-0d40b515d99b-kube-api-access-h4x8g\") pod \"a052dc18-c68b-4e41-8666-0d40b515d99b\" (UID: \"a052dc18-c68b-4e41-8666-0d40b515d99b\") " Feb 25 13:32:04 crc kubenswrapper[4815]: I0225 13:32:04.950857 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a052dc18-c68b-4e41-8666-0d40b515d99b-kube-api-access-h4x8g" (OuterVolumeSpecName: "kube-api-access-h4x8g") pod "a052dc18-c68b-4e41-8666-0d40b515d99b" (UID: "a052dc18-c68b-4e41-8666-0d40b515d99b"). InnerVolumeSpecName "kube-api-access-h4x8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:32:05 crc kubenswrapper[4815]: I0225 13:32:05.044942 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4x8g\" (UniqueName: \"kubernetes.io/projected/a052dc18-c68b-4e41-8666-0d40b515d99b-kube-api-access-h4x8g\") on node \"crc\" DevicePath \"\"" Feb 25 13:32:05 crc kubenswrapper[4815]: I0225 13:32:05.515442 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533772-htrc4" event={"ID":"a052dc18-c68b-4e41-8666-0d40b515d99b","Type":"ContainerDied","Data":"faa3a10092648070370243fec3b647fe7143b836c0966d5e606b160de289b433"} Feb 25 13:32:05 crc kubenswrapper[4815]: I0225 13:32:05.515501 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="faa3a10092648070370243fec3b647fe7143b836c0966d5e606b160de289b433" Feb 25 13:32:05 crc kubenswrapper[4815]: I0225 13:32:05.515781 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533772-htrc4" Feb 25 13:32:05 crc kubenswrapper[4815]: I0225 13:32:05.564320 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533766-d62n6"] Feb 25 13:32:05 crc kubenswrapper[4815]: I0225 13:32:05.568688 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533766-d62n6"] Feb 25 13:32:06 crc kubenswrapper[4815]: I0225 13:32:06.943248 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bfd8350-d886-4279-aac6-28e1f1c1f752" path="/var/lib/kubelet/pods/5bfd8350-d886-4279-aac6-28e1f1c1f752/volumes" Feb 25 13:32:07 crc kubenswrapper[4815]: I0225 13:32:07.626588 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-sgk4q" Feb 25 13:32:20 crc kubenswrapper[4815]: I0225 13:32:20.943664 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx"] Feb 25 13:32:20 crc kubenswrapper[4815]: E0225 13:32:20.944446 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a052dc18-c68b-4e41-8666-0d40b515d99b" containerName="oc" Feb 25 13:32:20 crc kubenswrapper[4815]: I0225 13:32:20.944463 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a052dc18-c68b-4e41-8666-0d40b515d99b" containerName="oc" Feb 25 13:32:20 crc kubenswrapper[4815]: I0225 13:32:20.944616 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="a052dc18-c68b-4e41-8666-0d40b515d99b" containerName="oc" Feb 25 13:32:20 crc kubenswrapper[4815]: I0225 13:32:20.945985 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" Feb 25 13:32:20 crc kubenswrapper[4815]: I0225 13:32:20.949242 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 25 13:32:20 crc kubenswrapper[4815]: I0225 13:32:20.950870 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx"] Feb 25 13:32:21 crc kubenswrapper[4815]: I0225 13:32:21.082964 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rrjl\" (UniqueName: \"kubernetes.io/projected/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-kube-api-access-9rrjl\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx\" (UID: \"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" Feb 25 13:32:21 crc kubenswrapper[4815]: I0225 13:32:21.083227 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx\" (UID: \"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" Feb 25 13:32:21 crc kubenswrapper[4815]: I0225 13:32:21.083295 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx\" (UID: \"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" Feb 25 13:32:21 crc kubenswrapper[4815]: I0225 13:32:21.184469 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx\" (UID: \"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" Feb 25 13:32:21 crc kubenswrapper[4815]: I0225 13:32:21.184586 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rrjl\" (UniqueName: \"kubernetes.io/projected/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-kube-api-access-9rrjl\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx\" (UID: \"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" Feb 25 13:32:21 crc kubenswrapper[4815]: I0225 13:32:21.184631 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx\" (UID: \"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" Feb 25 13:32:21 crc kubenswrapper[4815]: I0225 13:32:21.185280 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx\" (UID: \"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" Feb 25 13:32:21 crc kubenswrapper[4815]: I0225 13:32:21.185303 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx\" (UID: \"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" Feb 25 13:32:21 crc kubenswrapper[4815]: I0225 13:32:21.212249 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rrjl\" (UniqueName: \"kubernetes.io/projected/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-kube-api-access-9rrjl\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx\" (UID: \"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" Feb 25 13:32:21 crc kubenswrapper[4815]: I0225 13:32:21.262027 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" Feb 25 13:32:21 crc kubenswrapper[4815]: I0225 13:32:21.551516 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx"] Feb 25 13:32:21 crc kubenswrapper[4815]: I0225 13:32:21.636308 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" event={"ID":"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5","Type":"ContainerStarted","Data":"d4b5cb7cedb56847589372b7164139d4889b5a933978df6bcd27c8bb6c04ccd9"} Feb 25 13:32:22 crc kubenswrapper[4815]: I0225 13:32:22.645853 4815 generic.go:334] "Generic (PLEG): container finished" podID="8a6513b5-471c-4fb3-bdd7-4cf5e1528db5" containerID="3faeecc7fa86abc4e0d8bddc3d83d874ed2a2a8b77405dbae5be3d095a91b0b2" exitCode=0 Feb 25 13:32:22 crc kubenswrapper[4815]: I0225 13:32:22.645908 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" event={"ID":"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5","Type":"ContainerDied","Data":"3faeecc7fa86abc4e0d8bddc3d83d874ed2a2a8b77405dbae5be3d095a91b0b2"} Feb 25 13:32:23 crc kubenswrapper[4815]: I0225 13:32:23.592703 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-mhgxd" podUID="a81edad3-179a-4786-9ae9-e1a8f0a4d01a" containerName="console" containerID="cri-o://42fd30cefd32dca44fba88b4d91e294bcfe2c963e00cd0a323bb898e8e032242" gracePeriod=15 Feb 25 13:32:23 crc kubenswrapper[4815]: I0225 13:32:23.962914 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-mhgxd_a81edad3-179a-4786-9ae9-e1a8f0a4d01a/console/0.log" Feb 25 13:32:23 crc kubenswrapper[4815]: I0225 13:32:23.963006 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.046935 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-config\") pod \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.047002 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-trusted-ca-bundle\") pod \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.047087 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-serving-cert\") pod \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.047178 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-oauth-serving-cert\") pod \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.048232 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "a81edad3-179a-4786-9ae9-e1a8f0a4d01a" (UID: "a81edad3-179a-4786-9ae9-e1a8f0a4d01a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.048245 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-config" (OuterVolumeSpecName: "console-config") pod "a81edad3-179a-4786-9ae9-e1a8f0a4d01a" (UID: "a81edad3-179a-4786-9ae9-e1a8f0a4d01a"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.048920 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "a81edad3-179a-4786-9ae9-e1a8f0a4d01a" (UID: "a81edad3-179a-4786-9ae9-e1a8f0a4d01a"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.049005 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-service-ca\") pod \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.049044 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6zdr\" (UniqueName: \"kubernetes.io/projected/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-kube-api-access-c6zdr\") pod \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.049082 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-oauth-config\") pod \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\" (UID: \"a81edad3-179a-4786-9ae9-e1a8f0a4d01a\") " Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.050246 4815 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.050308 4815 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.050326 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.050336 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-service-ca" (OuterVolumeSpecName: "service-ca") pod "a81edad3-179a-4786-9ae9-e1a8f0a4d01a" (UID: "a81edad3-179a-4786-9ae9-e1a8f0a4d01a"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.054904 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "a81edad3-179a-4786-9ae9-e1a8f0a4d01a" (UID: "a81edad3-179a-4786-9ae9-e1a8f0a4d01a"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.055302 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "a81edad3-179a-4786-9ae9-e1a8f0a4d01a" (UID: "a81edad3-179a-4786-9ae9-e1a8f0a4d01a"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.055802 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-kube-api-access-c6zdr" (OuterVolumeSpecName: "kube-api-access-c6zdr") pod "a81edad3-179a-4786-9ae9-e1a8f0a4d01a" (UID: "a81edad3-179a-4786-9ae9-e1a8f0a4d01a"). InnerVolumeSpecName "kube-api-access-c6zdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.151763 4815 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.151802 4815 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-service-ca\") on node \"crc\" DevicePath \"\"" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.151812 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6zdr\" (UniqueName: \"kubernetes.io/projected/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-kube-api-access-c6zdr\") on node \"crc\" DevicePath \"\"" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.151822 4815 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a81edad3-179a-4786-9ae9-e1a8f0a4d01a-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.658824 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-mhgxd_a81edad3-179a-4786-9ae9-e1a8f0a4d01a/console/0.log" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.659063 4815 generic.go:334] "Generic (PLEG): container finished" podID="a81edad3-179a-4786-9ae9-e1a8f0a4d01a" containerID="42fd30cefd32dca44fba88b4d91e294bcfe2c963e00cd0a323bb898e8e032242" exitCode=2 Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.659165 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mhgxd" event={"ID":"a81edad3-179a-4786-9ae9-e1a8f0a4d01a","Type":"ContainerDied","Data":"42fd30cefd32dca44fba88b4d91e294bcfe2c963e00cd0a323bb898e8e032242"} Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.659239 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mhgxd" event={"ID":"a81edad3-179a-4786-9ae9-e1a8f0a4d01a","Type":"ContainerDied","Data":"0089f754c2c276e7a440758d91f196d27023f0649f2625a873528e9077ce8a42"} Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.659300 4815 scope.go:117] "RemoveContainer" containerID="42fd30cefd32dca44fba88b4d91e294bcfe2c963e00cd0a323bb898e8e032242" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.659426 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mhgxd" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.662960 4815 generic.go:334] "Generic (PLEG): container finished" podID="8a6513b5-471c-4fb3-bdd7-4cf5e1528db5" containerID="b25c7c2b17cbe91229b4ce0b92c8b4c29fde41f34901956566528317cda60912" exitCode=0 Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.663043 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" event={"ID":"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5","Type":"ContainerDied","Data":"b25c7c2b17cbe91229b4ce0b92c8b4c29fde41f34901956566528317cda60912"} Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.702492 4815 scope.go:117] "RemoveContainer" containerID="42fd30cefd32dca44fba88b4d91e294bcfe2c963e00cd0a323bb898e8e032242" Feb 25 13:32:24 crc kubenswrapper[4815]: E0225 13:32:24.704229 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42fd30cefd32dca44fba88b4d91e294bcfe2c963e00cd0a323bb898e8e032242\": container with ID starting with 42fd30cefd32dca44fba88b4d91e294bcfe2c963e00cd0a323bb898e8e032242 not found: ID does not exist" containerID="42fd30cefd32dca44fba88b4d91e294bcfe2c963e00cd0a323bb898e8e032242" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.704295 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42fd30cefd32dca44fba88b4d91e294bcfe2c963e00cd0a323bb898e8e032242"} err="failed to get container status \"42fd30cefd32dca44fba88b4d91e294bcfe2c963e00cd0a323bb898e8e032242\": rpc error: code = NotFound desc = could not find container \"42fd30cefd32dca44fba88b4d91e294bcfe2c963e00cd0a323bb898e8e032242\": container with ID starting with 42fd30cefd32dca44fba88b4d91e294bcfe2c963e00cd0a323bb898e8e032242 not found: ID does not exist" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.712399 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-mhgxd"] Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.717962 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-mhgxd"] Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.885788 4815 patch_prober.go:28] interesting pod/console-f9d7485db-mhgxd container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.885864 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-f9d7485db-mhgxd" podUID="a81edad3-179a-4786-9ae9-e1a8f0a4d01a" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 25 13:32:24 crc kubenswrapper[4815]: I0225 13:32:24.948502 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a81edad3-179a-4786-9ae9-e1a8f0a4d01a" path="/var/lib/kubelet/pods/a81edad3-179a-4786-9ae9-e1a8f0a4d01a/volumes" Feb 25 13:32:25 crc kubenswrapper[4815]: I0225 13:32:25.672134 4815 generic.go:334] "Generic (PLEG): container finished" podID="8a6513b5-471c-4fb3-bdd7-4cf5e1528db5" containerID="dedc919d6d04d7104635a91ea7369d7a8375a1e9a578298aa1f7119369ae4055" exitCode=0 Feb 25 13:32:25 crc kubenswrapper[4815]: I0225 13:32:25.672174 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" event={"ID":"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5","Type":"ContainerDied","Data":"dedc919d6d04d7104635a91ea7369d7a8375a1e9a578298aa1f7119369ae4055"} Feb 25 13:32:26 crc kubenswrapper[4815]: I0225 13:32:26.950052 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" Feb 25 13:32:26 crc kubenswrapper[4815]: I0225 13:32:26.992328 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-bundle\") pod \"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5\" (UID: \"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5\") " Feb 25 13:32:26 crc kubenswrapper[4815]: I0225 13:32:26.992403 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rrjl\" (UniqueName: \"kubernetes.io/projected/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-kube-api-access-9rrjl\") pod \"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5\" (UID: \"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5\") " Feb 25 13:32:26 crc kubenswrapper[4815]: I0225 13:32:26.992457 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-util\") pod \"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5\" (UID: \"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5\") " Feb 25 13:32:26 crc kubenswrapper[4815]: I0225 13:32:26.993440 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-bundle" (OuterVolumeSpecName: "bundle") pod "8a6513b5-471c-4fb3-bdd7-4cf5e1528db5" (UID: "8a6513b5-471c-4fb3-bdd7-4cf5e1528db5"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:32:26 crc kubenswrapper[4815]: I0225 13:32:26.999319 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-kube-api-access-9rrjl" (OuterVolumeSpecName: "kube-api-access-9rrjl") pod "8a6513b5-471c-4fb3-bdd7-4cf5e1528db5" (UID: "8a6513b5-471c-4fb3-bdd7-4cf5e1528db5"). InnerVolumeSpecName "kube-api-access-9rrjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:32:27 crc kubenswrapper[4815]: I0225 13:32:27.022129 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-util" (OuterVolumeSpecName: "util") pod "8a6513b5-471c-4fb3-bdd7-4cf5e1528db5" (UID: "8a6513b5-471c-4fb3-bdd7-4cf5e1528db5"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:32:27 crc kubenswrapper[4815]: I0225 13:32:27.093752 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rrjl\" (UniqueName: \"kubernetes.io/projected/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-kube-api-access-9rrjl\") on node \"crc\" DevicePath \"\"" Feb 25 13:32:27 crc kubenswrapper[4815]: I0225 13:32:27.094029 4815 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-util\") on node \"crc\" DevicePath \"\"" Feb 25 13:32:27 crc kubenswrapper[4815]: I0225 13:32:27.094042 4815 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8a6513b5-471c-4fb3-bdd7-4cf5e1528db5-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:32:27 crc kubenswrapper[4815]: I0225 13:32:27.688192 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" event={"ID":"8a6513b5-471c-4fb3-bdd7-4cf5e1528db5","Type":"ContainerDied","Data":"d4b5cb7cedb56847589372b7164139d4889b5a933978df6bcd27c8bb6c04ccd9"} Feb 25 13:32:27 crc kubenswrapper[4815]: I0225 13:32:27.688253 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4b5cb7cedb56847589372b7164139d4889b5a933978df6bcd27c8bb6c04ccd9" Feb 25 13:32:27 crc kubenswrapper[4815]: I0225 13:32:27.688277 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx" Feb 25 13:32:35 crc kubenswrapper[4815]: I0225 13:32:35.941567 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv"] Feb 25 13:32:35 crc kubenswrapper[4815]: E0225 13:32:35.942205 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6513b5-471c-4fb3-bdd7-4cf5e1528db5" containerName="util" Feb 25 13:32:35 crc kubenswrapper[4815]: I0225 13:32:35.942216 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6513b5-471c-4fb3-bdd7-4cf5e1528db5" containerName="util" Feb 25 13:32:35 crc kubenswrapper[4815]: E0225 13:32:35.942228 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a81edad3-179a-4786-9ae9-e1a8f0a4d01a" containerName="console" Feb 25 13:32:35 crc kubenswrapper[4815]: I0225 13:32:35.942234 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a81edad3-179a-4786-9ae9-e1a8f0a4d01a" containerName="console" Feb 25 13:32:35 crc kubenswrapper[4815]: E0225 13:32:35.942246 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6513b5-471c-4fb3-bdd7-4cf5e1528db5" containerName="extract" Feb 25 13:32:35 crc kubenswrapper[4815]: I0225 13:32:35.942252 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6513b5-471c-4fb3-bdd7-4cf5e1528db5" containerName="extract" Feb 25 13:32:35 crc kubenswrapper[4815]: E0225 13:32:35.942265 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6513b5-471c-4fb3-bdd7-4cf5e1528db5" containerName="pull" Feb 25 13:32:35 crc kubenswrapper[4815]: I0225 13:32:35.942270 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6513b5-471c-4fb3-bdd7-4cf5e1528db5" containerName="pull" Feb 25 13:32:35 crc kubenswrapper[4815]: I0225 13:32:35.942360 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="a81edad3-179a-4786-9ae9-e1a8f0a4d01a" containerName="console" Feb 25 13:32:35 crc kubenswrapper[4815]: I0225 13:32:35.942370 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a6513b5-471c-4fb3-bdd7-4cf5e1528db5" containerName="extract" Feb 25 13:32:35 crc kubenswrapper[4815]: I0225 13:32:35.942731 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv" Feb 25 13:32:35 crc kubenswrapper[4815]: I0225 13:32:35.945492 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-q6ztp" Feb 25 13:32:35 crc kubenswrapper[4815]: I0225 13:32:35.945856 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 25 13:32:35 crc kubenswrapper[4815]: I0225 13:32:35.945881 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 25 13:32:35 crc kubenswrapper[4815]: I0225 13:32:35.945928 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 25 13:32:35 crc kubenswrapper[4815]: I0225 13:32:35.946435 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 25 13:32:35 crc kubenswrapper[4815]: I0225 13:32:35.964810 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv"] Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.021090 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cqhg\" (UniqueName: \"kubernetes.io/projected/e0c4f375-94ec-47b4-885c-0b86b980580e-kube-api-access-7cqhg\") pod \"metallb-operator-controller-manager-8674bc4d66-cm4rv\" (UID: \"e0c4f375-94ec-47b4-885c-0b86b980580e\") " pod="metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.021171 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e0c4f375-94ec-47b4-885c-0b86b980580e-apiservice-cert\") pod \"metallb-operator-controller-manager-8674bc4d66-cm4rv\" (UID: \"e0c4f375-94ec-47b4-885c-0b86b980580e\") " pod="metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.021193 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e0c4f375-94ec-47b4-885c-0b86b980580e-webhook-cert\") pod \"metallb-operator-controller-manager-8674bc4d66-cm4rv\" (UID: \"e0c4f375-94ec-47b4-885c-0b86b980580e\") " pod="metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.122405 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e0c4f375-94ec-47b4-885c-0b86b980580e-apiservice-cert\") pod \"metallb-operator-controller-manager-8674bc4d66-cm4rv\" (UID: \"e0c4f375-94ec-47b4-885c-0b86b980580e\") " pod="metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.122464 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e0c4f375-94ec-47b4-885c-0b86b980580e-webhook-cert\") pod \"metallb-operator-controller-manager-8674bc4d66-cm4rv\" (UID: \"e0c4f375-94ec-47b4-885c-0b86b980580e\") " pod="metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.122537 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cqhg\" (UniqueName: \"kubernetes.io/projected/e0c4f375-94ec-47b4-885c-0b86b980580e-kube-api-access-7cqhg\") pod \"metallb-operator-controller-manager-8674bc4d66-cm4rv\" (UID: \"e0c4f375-94ec-47b4-885c-0b86b980580e\") " pod="metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.140323 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e0c4f375-94ec-47b4-885c-0b86b980580e-apiservice-cert\") pod \"metallb-operator-controller-manager-8674bc4d66-cm4rv\" (UID: \"e0c4f375-94ec-47b4-885c-0b86b980580e\") " pod="metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.140389 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e0c4f375-94ec-47b4-885c-0b86b980580e-webhook-cert\") pod \"metallb-operator-controller-manager-8674bc4d66-cm4rv\" (UID: \"e0c4f375-94ec-47b4-885c-0b86b980580e\") " pod="metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.143853 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cqhg\" (UniqueName: \"kubernetes.io/projected/e0c4f375-94ec-47b4-885c-0b86b980580e-kube-api-access-7cqhg\") pod \"metallb-operator-controller-manager-8674bc4d66-cm4rv\" (UID: \"e0c4f375-94ec-47b4-885c-0b86b980580e\") " pod="metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.235589 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b"] Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.236442 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.238326 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-hgrs6" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.238686 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.238833 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.249282 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b"] Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.256472 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.327217 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j4ds\" (UniqueName: \"kubernetes.io/projected/2f7b411f-fe90-48bf-995b-a40a6a8357ee-kube-api-access-5j4ds\") pod \"metallb-operator-webhook-server-5b77f59548-v7w5b\" (UID: \"2f7b411f-fe90-48bf-995b-a40a6a8357ee\") " pod="metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.327616 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2f7b411f-fe90-48bf-995b-a40a6a8357ee-apiservice-cert\") pod \"metallb-operator-webhook-server-5b77f59548-v7w5b\" (UID: \"2f7b411f-fe90-48bf-995b-a40a6a8357ee\") " pod="metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.327657 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2f7b411f-fe90-48bf-995b-a40a6a8357ee-webhook-cert\") pod \"metallb-operator-webhook-server-5b77f59548-v7w5b\" (UID: \"2f7b411f-fe90-48bf-995b-a40a6a8357ee\") " pod="metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.429254 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j4ds\" (UniqueName: \"kubernetes.io/projected/2f7b411f-fe90-48bf-995b-a40a6a8357ee-kube-api-access-5j4ds\") pod \"metallb-operator-webhook-server-5b77f59548-v7w5b\" (UID: \"2f7b411f-fe90-48bf-995b-a40a6a8357ee\") " pod="metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.429353 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2f7b411f-fe90-48bf-995b-a40a6a8357ee-apiservice-cert\") pod \"metallb-operator-webhook-server-5b77f59548-v7w5b\" (UID: \"2f7b411f-fe90-48bf-995b-a40a6a8357ee\") " pod="metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.429391 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2f7b411f-fe90-48bf-995b-a40a6a8357ee-webhook-cert\") pod \"metallb-operator-webhook-server-5b77f59548-v7w5b\" (UID: \"2f7b411f-fe90-48bf-995b-a40a6a8357ee\") " pod="metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.433728 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2f7b411f-fe90-48bf-995b-a40a6a8357ee-apiservice-cert\") pod \"metallb-operator-webhook-server-5b77f59548-v7w5b\" (UID: \"2f7b411f-fe90-48bf-995b-a40a6a8357ee\") " pod="metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.454243 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2f7b411f-fe90-48bf-995b-a40a6a8357ee-webhook-cert\") pod \"metallb-operator-webhook-server-5b77f59548-v7w5b\" (UID: \"2f7b411f-fe90-48bf-995b-a40a6a8357ee\") " pod="metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.472300 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j4ds\" (UniqueName: \"kubernetes.io/projected/2f7b411f-fe90-48bf-995b-a40a6a8357ee-kube-api-access-5j4ds\") pod \"metallb-operator-webhook-server-5b77f59548-v7w5b\" (UID: \"2f7b411f-fe90-48bf-995b-a40a6a8357ee\") " pod="metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.561756 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b" Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.621534 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv"] Feb 25 13:32:36 crc kubenswrapper[4815]: W0225 13:32:36.632241 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0c4f375_94ec_47b4_885c_0b86b980580e.slice/crio-06fee3ecb05e0b2677a5e0103ffa1091fbdd6a5c9016615f2cea7e2efbb9b9d2 WatchSource:0}: Error finding container 06fee3ecb05e0b2677a5e0103ffa1091fbdd6a5c9016615f2cea7e2efbb9b9d2: Status 404 returned error can't find the container with id 06fee3ecb05e0b2677a5e0103ffa1091fbdd6a5c9016615f2cea7e2efbb9b9d2 Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.743574 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv" event={"ID":"e0c4f375-94ec-47b4-885c-0b86b980580e","Type":"ContainerStarted","Data":"06fee3ecb05e0b2677a5e0103ffa1091fbdd6a5c9016615f2cea7e2efbb9b9d2"} Feb 25 13:32:36 crc kubenswrapper[4815]: I0225 13:32:36.777711 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b"] Feb 25 13:32:36 crc kubenswrapper[4815]: W0225 13:32:36.790017 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f7b411f_fe90_48bf_995b_a40a6a8357ee.slice/crio-a2244e90ed1f560e56e3ee9ba6dfea3f2e6405df9039a02a678f2e50b1eb163b WatchSource:0}: Error finding container a2244e90ed1f560e56e3ee9ba6dfea3f2e6405df9039a02a678f2e50b1eb163b: Status 404 returned error can't find the container with id a2244e90ed1f560e56e3ee9ba6dfea3f2e6405df9039a02a678f2e50b1eb163b Feb 25 13:32:37 crc kubenswrapper[4815]: I0225 13:32:37.751097 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b" event={"ID":"2f7b411f-fe90-48bf-995b-a40a6a8357ee","Type":"ContainerStarted","Data":"a2244e90ed1f560e56e3ee9ba6dfea3f2e6405df9039a02a678f2e50b1eb163b"} Feb 25 13:32:41 crc kubenswrapper[4815]: I0225 13:32:41.778435 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b" event={"ID":"2f7b411f-fe90-48bf-995b-a40a6a8357ee","Type":"ContainerStarted","Data":"8fe46efe6701f516b7450b316534511420a3613cd2ffc23e72a60c68bd1624c7"} Feb 25 13:32:41 crc kubenswrapper[4815]: I0225 13:32:41.778971 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b" Feb 25 13:32:41 crc kubenswrapper[4815]: I0225 13:32:41.780474 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv" event={"ID":"e0c4f375-94ec-47b4-885c-0b86b980580e","Type":"ContainerStarted","Data":"b25c48dfece3e13c166d6948cff4b24bc90863f4321eadfb1f4349da4335c2eb"} Feb 25 13:32:41 crc kubenswrapper[4815]: I0225 13:32:41.780654 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv" Feb 25 13:32:41 crc kubenswrapper[4815]: I0225 13:32:41.800076 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b" podStartSLOduration=1.4820181350000001 podStartE2EDuration="5.800052714s" podCreationTimestamp="2026-02-25 13:32:36 +0000 UTC" firstStartedPulling="2026-02-25 13:32:36.793617055 +0000 UTC m=+714.594715109" lastFinishedPulling="2026-02-25 13:32:41.111651634 +0000 UTC m=+718.912749688" observedRunningTime="2026-02-25 13:32:41.795040604 +0000 UTC m=+719.596138678" watchObservedRunningTime="2026-02-25 13:32:41.800052714 +0000 UTC m=+719.601150768" Feb 25 13:32:41 crc kubenswrapper[4815]: I0225 13:32:41.810308 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv" podStartSLOduration=2.339622183 podStartE2EDuration="6.810288079s" podCreationTimestamp="2026-02-25 13:32:35 +0000 UTC" firstStartedPulling="2026-02-25 13:32:36.636592279 +0000 UTC m=+714.437690333" lastFinishedPulling="2026-02-25 13:32:41.107258175 +0000 UTC m=+718.908356229" observedRunningTime="2026-02-25 13:32:41.809274546 +0000 UTC m=+719.610372600" watchObservedRunningTime="2026-02-25 13:32:41.810288079 +0000 UTC m=+719.611386143" Feb 25 13:32:43 crc kubenswrapper[4815]: I0225 13:32:43.456188 4815 scope.go:117] "RemoveContainer" containerID="e8da71b1f37781e767db03771dca29938182875e21753e4c582698e896a74a8a" Feb 25 13:32:44 crc kubenswrapper[4815]: I0225 13:32:44.432975 4815 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 25 13:32:56 crc kubenswrapper[4815]: I0225 13:32:56.565431 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5b77f59548-v7w5b" Feb 25 13:33:16 crc kubenswrapper[4815]: I0225 13:33:16.260553 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-8674bc4d66-cm4rv" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.700192 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-92bcq"] Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.702944 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.704125 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-kbf7c" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.704732 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.704742 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.705382 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-mcphd"] Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.706224 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-mcphd" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.709085 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.716003 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-mcphd"] Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.791120 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-qxvv8"] Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.791981 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qxvv8" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.794222 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.794270 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.794379 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.794611 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-gwzh5" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.805056 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-69bbfbf88f-f6fqm"] Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.805968 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-f6fqm" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.807389 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.818425 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-f6fqm"] Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.842098 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/801eb953-96a0-4272-9f81-13e80e61f75a-frr-startup\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.842140 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/801eb953-96a0-4272-9f81-13e80e61f75a-frr-conf\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.842162 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-862w7\" (UniqueName: \"kubernetes.io/projected/699a6712-3fed-4b60-a2ea-e7440dd0e306-kube-api-access-862w7\") pod \"frr-k8s-webhook-server-78b44bf5bb-mcphd\" (UID: \"699a6712-3fed-4b60-a2ea-e7440dd0e306\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-mcphd" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.842185 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/801eb953-96a0-4272-9f81-13e80e61f75a-frr-sockets\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.842306 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mfnl\" (UniqueName: \"kubernetes.io/projected/801eb953-96a0-4272-9f81-13e80e61f75a-kube-api-access-4mfnl\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.842434 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/801eb953-96a0-4272-9f81-13e80e61f75a-reloader\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.842545 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/699a6712-3fed-4b60-a2ea-e7440dd0e306-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-mcphd\" (UID: \"699a6712-3fed-4b60-a2ea-e7440dd0e306\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-mcphd" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.842602 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/801eb953-96a0-4272-9f81-13e80e61f75a-metrics\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.842645 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/801eb953-96a0-4272-9f81-13e80e61f75a-metrics-certs\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.944278 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/801eb953-96a0-4272-9f81-13e80e61f75a-metrics\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.944362 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cefeaa51-3ff5-4124-8e3d-879d579f0d91-metrics-certs\") pod \"speaker-qxvv8\" (UID: \"cefeaa51-3ff5-4124-8e3d-879d579f0d91\") " pod="metallb-system/speaker-qxvv8" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.944414 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/801eb953-96a0-4272-9f81-13e80e61f75a-metrics-certs\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.944443 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1b6c9556-1980-4581-8de1-f1fbf5d8256a-metrics-certs\") pod \"controller-69bbfbf88f-f6fqm\" (UID: \"1b6c9556-1980-4581-8de1-f1fbf5d8256a\") " pod="metallb-system/controller-69bbfbf88f-f6fqm" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.944537 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/801eb953-96a0-4272-9f81-13e80e61f75a-frr-startup\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.944560 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqd59\" (UniqueName: \"kubernetes.io/projected/cefeaa51-3ff5-4124-8e3d-879d579f0d91-kube-api-access-dqd59\") pod \"speaker-qxvv8\" (UID: \"cefeaa51-3ff5-4124-8e3d-879d579f0d91\") " pod="metallb-system/speaker-qxvv8" Feb 25 13:33:17 crc kubenswrapper[4815]: E0225 13:33:17.944571 4815 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.944582 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/801eb953-96a0-4272-9f81-13e80e61f75a-frr-conf\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: E0225 13:33:17.944640 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/801eb953-96a0-4272-9f81-13e80e61f75a-metrics-certs podName:801eb953-96a0-4272-9f81-13e80e61f75a nodeName:}" failed. No retries permitted until 2026-02-25 13:33:18.44462084 +0000 UTC m=+756.245718884 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/801eb953-96a0-4272-9f81-13e80e61f75a-metrics-certs") pod "frr-k8s-92bcq" (UID: "801eb953-96a0-4272-9f81-13e80e61f75a") : secret "frr-k8s-certs-secret" not found Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.944664 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-862w7\" (UniqueName: \"kubernetes.io/projected/699a6712-3fed-4b60-a2ea-e7440dd0e306-kube-api-access-862w7\") pod \"frr-k8s-webhook-server-78b44bf5bb-mcphd\" (UID: \"699a6712-3fed-4b60-a2ea-e7440dd0e306\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-mcphd" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.944694 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/801eb953-96a0-4272-9f81-13e80e61f75a-frr-sockets\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.944734 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b6c9556-1980-4581-8de1-f1fbf5d8256a-cert\") pod \"controller-69bbfbf88f-f6fqm\" (UID: \"1b6c9556-1980-4581-8de1-f1fbf5d8256a\") " pod="metallb-system/controller-69bbfbf88f-f6fqm" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.944761 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mfnl\" (UniqueName: \"kubernetes.io/projected/801eb953-96a0-4272-9f81-13e80e61f75a-kube-api-access-4mfnl\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.944814 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/cefeaa51-3ff5-4124-8e3d-879d579f0d91-memberlist\") pod \"speaker-qxvv8\" (UID: \"cefeaa51-3ff5-4124-8e3d-879d579f0d91\") " pod="metallb-system/speaker-qxvv8" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.944837 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/801eb953-96a0-4272-9f81-13e80e61f75a-reloader\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.944887 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/699a6712-3fed-4b60-a2ea-e7440dd0e306-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-mcphd\" (UID: \"699a6712-3fed-4b60-a2ea-e7440dd0e306\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-mcphd" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.944912 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8nht\" (UniqueName: \"kubernetes.io/projected/1b6c9556-1980-4581-8de1-f1fbf5d8256a-kube-api-access-t8nht\") pod \"controller-69bbfbf88f-f6fqm\" (UID: \"1b6c9556-1980-4581-8de1-f1fbf5d8256a\") " pod="metallb-system/controller-69bbfbf88f-f6fqm" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.944929 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/cefeaa51-3ff5-4124-8e3d-879d579f0d91-metallb-excludel2\") pod \"speaker-qxvv8\" (UID: \"cefeaa51-3ff5-4124-8e3d-879d579f0d91\") " pod="metallb-system/speaker-qxvv8" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.944984 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/801eb953-96a0-4272-9f81-13e80e61f75a-metrics\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.945268 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/801eb953-96a0-4272-9f81-13e80e61f75a-reloader\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.945465 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/801eb953-96a0-4272-9f81-13e80e61f75a-frr-sockets\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.945478 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/801eb953-96a0-4272-9f81-13e80e61f75a-frr-startup\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.945607 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/801eb953-96a0-4272-9f81-13e80e61f75a-frr-conf\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.954368 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/699a6712-3fed-4b60-a2ea-e7440dd0e306-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-mcphd\" (UID: \"699a6712-3fed-4b60-a2ea-e7440dd0e306\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-mcphd" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.963260 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-862w7\" (UniqueName: \"kubernetes.io/projected/699a6712-3fed-4b60-a2ea-e7440dd0e306-kube-api-access-862w7\") pod \"frr-k8s-webhook-server-78b44bf5bb-mcphd\" (UID: \"699a6712-3fed-4b60-a2ea-e7440dd0e306\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-mcphd" Feb 25 13:33:17 crc kubenswrapper[4815]: I0225 13:33:17.970195 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mfnl\" (UniqueName: \"kubernetes.io/projected/801eb953-96a0-4272-9f81-13e80e61f75a-kube-api-access-4mfnl\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.026490 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-mcphd" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.046238 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b6c9556-1980-4581-8de1-f1fbf5d8256a-cert\") pod \"controller-69bbfbf88f-f6fqm\" (UID: \"1b6c9556-1980-4581-8de1-f1fbf5d8256a\") " pod="metallb-system/controller-69bbfbf88f-f6fqm" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.046307 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/cefeaa51-3ff5-4124-8e3d-879d579f0d91-memberlist\") pod \"speaker-qxvv8\" (UID: \"cefeaa51-3ff5-4124-8e3d-879d579f0d91\") " pod="metallb-system/speaker-qxvv8" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.046358 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8nht\" (UniqueName: \"kubernetes.io/projected/1b6c9556-1980-4581-8de1-f1fbf5d8256a-kube-api-access-t8nht\") pod \"controller-69bbfbf88f-f6fqm\" (UID: \"1b6c9556-1980-4581-8de1-f1fbf5d8256a\") " pod="metallb-system/controller-69bbfbf88f-f6fqm" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.046408 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/cefeaa51-3ff5-4124-8e3d-879d579f0d91-metallb-excludel2\") pod \"speaker-qxvv8\" (UID: \"cefeaa51-3ff5-4124-8e3d-879d579f0d91\") " pod="metallb-system/speaker-qxvv8" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.046436 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cefeaa51-3ff5-4124-8e3d-879d579f0d91-metrics-certs\") pod \"speaker-qxvv8\" (UID: \"cefeaa51-3ff5-4124-8e3d-879d579f0d91\") " pod="metallb-system/speaker-qxvv8" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.046491 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1b6c9556-1980-4581-8de1-f1fbf5d8256a-metrics-certs\") pod \"controller-69bbfbf88f-f6fqm\" (UID: \"1b6c9556-1980-4581-8de1-f1fbf5d8256a\") " pod="metallb-system/controller-69bbfbf88f-f6fqm" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.046541 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqd59\" (UniqueName: \"kubernetes.io/projected/cefeaa51-3ff5-4124-8e3d-879d579f0d91-kube-api-access-dqd59\") pod \"speaker-qxvv8\" (UID: \"cefeaa51-3ff5-4124-8e3d-879d579f0d91\") " pod="metallb-system/speaker-qxvv8" Feb 25 13:33:18 crc kubenswrapper[4815]: E0225 13:33:18.046713 4815 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Feb 25 13:33:18 crc kubenswrapper[4815]: E0225 13:33:18.046771 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cefeaa51-3ff5-4124-8e3d-879d579f0d91-metrics-certs podName:cefeaa51-3ff5-4124-8e3d-879d579f0d91 nodeName:}" failed. No retries permitted until 2026-02-25 13:33:18.546752413 +0000 UTC m=+756.347850537 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cefeaa51-3ff5-4124-8e3d-879d579f0d91-metrics-certs") pod "speaker-qxvv8" (UID: "cefeaa51-3ff5-4124-8e3d-879d579f0d91") : secret "speaker-certs-secret" not found Feb 25 13:33:18 crc kubenswrapper[4815]: E0225 13:33:18.047167 4815 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 25 13:33:18 crc kubenswrapper[4815]: E0225 13:33:18.047249 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cefeaa51-3ff5-4124-8e3d-879d579f0d91-memberlist podName:cefeaa51-3ff5-4124-8e3d-879d579f0d91 nodeName:}" failed. No retries permitted until 2026-02-25 13:33:18.547231098 +0000 UTC m=+756.348329152 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/cefeaa51-3ff5-4124-8e3d-879d579f0d91-memberlist") pod "speaker-qxvv8" (UID: "cefeaa51-3ff5-4124-8e3d-879d579f0d91") : secret "metallb-memberlist" not found Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.047461 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/cefeaa51-3ff5-4124-8e3d-879d579f0d91-metallb-excludel2\") pod \"speaker-qxvv8\" (UID: \"cefeaa51-3ff5-4124-8e3d-879d579f0d91\") " pod="metallb-system/speaker-qxvv8" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.048746 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.050447 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1b6c9556-1980-4581-8de1-f1fbf5d8256a-metrics-certs\") pod \"controller-69bbfbf88f-f6fqm\" (UID: \"1b6c9556-1980-4581-8de1-f1fbf5d8256a\") " pod="metallb-system/controller-69bbfbf88f-f6fqm" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.061558 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b6c9556-1980-4581-8de1-f1fbf5d8256a-cert\") pod \"controller-69bbfbf88f-f6fqm\" (UID: \"1b6c9556-1980-4581-8de1-f1fbf5d8256a\") " pod="metallb-system/controller-69bbfbf88f-f6fqm" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.065907 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8nht\" (UniqueName: \"kubernetes.io/projected/1b6c9556-1980-4581-8de1-f1fbf5d8256a-kube-api-access-t8nht\") pod \"controller-69bbfbf88f-f6fqm\" (UID: \"1b6c9556-1980-4581-8de1-f1fbf5d8256a\") " pod="metallb-system/controller-69bbfbf88f-f6fqm" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.069093 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqd59\" (UniqueName: \"kubernetes.io/projected/cefeaa51-3ff5-4124-8e3d-879d579f0d91-kube-api-access-dqd59\") pod \"speaker-qxvv8\" (UID: \"cefeaa51-3ff5-4124-8e3d-879d579f0d91\") " pod="metallb-system/speaker-qxvv8" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.118017 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-f6fqm" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.268266 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-mcphd"] Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.274271 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.344979 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-f6fqm"] Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.451050 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/801eb953-96a0-4272-9f81-13e80e61f75a-metrics-certs\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.455776 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/801eb953-96a0-4272-9f81-13e80e61f75a-metrics-certs\") pod \"frr-k8s-92bcq\" (UID: \"801eb953-96a0-4272-9f81-13e80e61f75a\") " pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.551824 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/cefeaa51-3ff5-4124-8e3d-879d579f0d91-memberlist\") pod \"speaker-qxvv8\" (UID: \"cefeaa51-3ff5-4124-8e3d-879d579f0d91\") " pod="metallb-system/speaker-qxvv8" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.551886 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cefeaa51-3ff5-4124-8e3d-879d579f0d91-metrics-certs\") pod \"speaker-qxvv8\" (UID: \"cefeaa51-3ff5-4124-8e3d-879d579f0d91\") " pod="metallb-system/speaker-qxvv8" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.555651 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/cefeaa51-3ff5-4124-8e3d-879d579f0d91-memberlist\") pod \"speaker-qxvv8\" (UID: \"cefeaa51-3ff5-4124-8e3d-879d579f0d91\") " pod="metallb-system/speaker-qxvv8" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.559443 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cefeaa51-3ff5-4124-8e3d-879d579f0d91-metrics-certs\") pod \"speaker-qxvv8\" (UID: \"cefeaa51-3ff5-4124-8e3d-879d579f0d91\") " pod="metallb-system/speaker-qxvv8" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.619335 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.703796 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qxvv8" Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.996501 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qxvv8" event={"ID":"cefeaa51-3ff5-4124-8e3d-879d579f0d91","Type":"ContainerStarted","Data":"5f18846126715e2d02ee6ab7fc69d41eca3d1d68237d98dcda1e97e0fa3b24db"} Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.997568 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-92bcq" event={"ID":"801eb953-96a0-4272-9f81-13e80e61f75a","Type":"ContainerStarted","Data":"3fa13d3baa251e24bc32bd79a818f05ec69dd2c38d7a8ab75ec88610f84f4cec"} Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.999386 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-f6fqm" event={"ID":"1b6c9556-1980-4581-8de1-f1fbf5d8256a","Type":"ContainerStarted","Data":"74ab495fc69ac735b81e97bf894fa21ce493c1146ca60df90a12f1b25b1487f0"} Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.999458 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-f6fqm" event={"ID":"1b6c9556-1980-4581-8de1-f1fbf5d8256a","Type":"ContainerStarted","Data":"e5be9be8a756eda11a982727dca1ec3af7c17cc8cd5f36f149616c71a0d95db5"} Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.999470 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-f6fqm" event={"ID":"1b6c9556-1980-4581-8de1-f1fbf5d8256a","Type":"ContainerStarted","Data":"230ae822fde1344c9315fd93df39a40bbf2af7c2e01b0ef5081574e78a3561b5"} Feb 25 13:33:18 crc kubenswrapper[4815]: I0225 13:33:18.999491 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-69bbfbf88f-f6fqm" Feb 25 13:33:19 crc kubenswrapper[4815]: I0225 13:33:19.000496 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-mcphd" event={"ID":"699a6712-3fed-4b60-a2ea-e7440dd0e306","Type":"ContainerStarted","Data":"72dfc484655a895d8e8e2f8796ea2aad86c04ed7bf3437440d8d471691f96e4b"} Feb 25 13:33:19 crc kubenswrapper[4815]: I0225 13:33:19.017771 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-69bbfbf88f-f6fqm" podStartSLOduration=2.0177576249999998 podStartE2EDuration="2.017757625s" podCreationTimestamp="2026-02-25 13:33:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:33:19.016109033 +0000 UTC m=+756.817207087" watchObservedRunningTime="2026-02-25 13:33:19.017757625 +0000 UTC m=+756.818855679" Feb 25 13:33:20 crc kubenswrapper[4815]: I0225 13:33:20.014927 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qxvv8" event={"ID":"cefeaa51-3ff5-4124-8e3d-879d579f0d91","Type":"ContainerStarted","Data":"79e035bbffd3d71fd44b708a0f69d52bf7e9ae27856b367a4390dc9b36e4f0d8"} Feb 25 13:33:20 crc kubenswrapper[4815]: I0225 13:33:20.015277 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-qxvv8" Feb 25 13:33:20 crc kubenswrapper[4815]: I0225 13:33:20.015291 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qxvv8" event={"ID":"cefeaa51-3ff5-4124-8e3d-879d579f0d91","Type":"ContainerStarted","Data":"0b684fcfded0733e57085cbda611740ecc97a110f27087dd272c05f7389ca758"} Feb 25 13:33:20 crc kubenswrapper[4815]: I0225 13:33:20.046853 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-qxvv8" podStartSLOduration=3.046810011 podStartE2EDuration="3.046810011s" podCreationTimestamp="2026-02-25 13:33:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:33:20.04365768 +0000 UTC m=+757.844755734" watchObservedRunningTime="2026-02-25 13:33:20.046810011 +0000 UTC m=+757.847908075" Feb 25 13:33:21 crc kubenswrapper[4815]: I0225 13:33:21.342501 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:33:21 crc kubenswrapper[4815]: I0225 13:33:21.342877 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:33:26 crc kubenswrapper[4815]: I0225 13:33:26.084687 4815 generic.go:334] "Generic (PLEG): container finished" podID="801eb953-96a0-4272-9f81-13e80e61f75a" containerID="7ba6bde65316d847d9d82411847e8c4a24e4df69ba452771df9d9a6f18befda6" exitCode=0 Feb 25 13:33:26 crc kubenswrapper[4815]: I0225 13:33:26.084778 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-92bcq" event={"ID":"801eb953-96a0-4272-9f81-13e80e61f75a","Type":"ContainerDied","Data":"7ba6bde65316d847d9d82411847e8c4a24e4df69ba452771df9d9a6f18befda6"} Feb 25 13:33:26 crc kubenswrapper[4815]: I0225 13:33:26.087881 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-mcphd" event={"ID":"699a6712-3fed-4b60-a2ea-e7440dd0e306","Type":"ContainerStarted","Data":"abc061ca7c2cdf1cf693b8f979bf8ed87a73de680c4512c11f35ada20098bf8d"} Feb 25 13:33:26 crc kubenswrapper[4815]: I0225 13:33:26.088059 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-mcphd" Feb 25 13:33:27 crc kubenswrapper[4815]: I0225 13:33:27.094527 4815 generic.go:334] "Generic (PLEG): container finished" podID="801eb953-96a0-4272-9f81-13e80e61f75a" containerID="33efe71fbf0582bde35a947b3b5e764a20add5b1171dcf018a42020b09cd56dc" exitCode=0 Feb 25 13:33:27 crc kubenswrapper[4815]: I0225 13:33:27.094653 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-92bcq" event={"ID":"801eb953-96a0-4272-9f81-13e80e61f75a","Type":"ContainerDied","Data":"33efe71fbf0582bde35a947b3b5e764a20add5b1171dcf018a42020b09cd56dc"} Feb 25 13:33:27 crc kubenswrapper[4815]: I0225 13:33:27.131144 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-mcphd" podStartSLOduration=3.45210539 podStartE2EDuration="10.131127908s" podCreationTimestamp="2026-02-25 13:33:17 +0000 UTC" firstStartedPulling="2026-02-25 13:33:18.274020899 +0000 UTC m=+756.075118953" lastFinishedPulling="2026-02-25 13:33:24.953043427 +0000 UTC m=+762.754141471" observedRunningTime="2026-02-25 13:33:26.137244889 +0000 UTC m=+763.938342943" watchObservedRunningTime="2026-02-25 13:33:27.131127908 +0000 UTC m=+764.932225962" Feb 25 13:33:28 crc kubenswrapper[4815]: I0225 13:33:28.102743 4815 generic.go:334] "Generic (PLEG): container finished" podID="801eb953-96a0-4272-9f81-13e80e61f75a" containerID="0b27c7ccf8d0ced36550511344d3b10db9c0e832112ea5df7563a286059cc42b" exitCode=0 Feb 25 13:33:28 crc kubenswrapper[4815]: I0225 13:33:28.102844 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-92bcq" event={"ID":"801eb953-96a0-4272-9f81-13e80e61f75a","Type":"ContainerDied","Data":"0b27c7ccf8d0ced36550511344d3b10db9c0e832112ea5df7563a286059cc42b"} Feb 25 13:33:28 crc kubenswrapper[4815]: I0225 13:33:28.140144 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-69bbfbf88f-f6fqm" Feb 25 13:33:29 crc kubenswrapper[4815]: I0225 13:33:29.113155 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-92bcq" event={"ID":"801eb953-96a0-4272-9f81-13e80e61f75a","Type":"ContainerStarted","Data":"4096961b5a21380a9378c60c60e66d655b5e37178b06e4f1e76b90cf736246f1"} Feb 25 13:33:29 crc kubenswrapper[4815]: I0225 13:33:29.113427 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-92bcq" event={"ID":"801eb953-96a0-4272-9f81-13e80e61f75a","Type":"ContainerStarted","Data":"d683f97280247bb80f54c18e14f4a70bea989c56e8b86641c733c2027bb9cee0"} Feb 25 13:33:29 crc kubenswrapper[4815]: I0225 13:33:29.113457 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:29 crc kubenswrapper[4815]: I0225 13:33:29.113471 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-92bcq" event={"ID":"801eb953-96a0-4272-9f81-13e80e61f75a","Type":"ContainerStarted","Data":"57c054ed743149d80190b70f0c1a5d0514469abed68676b2e3993e91202b53e8"} Feb 25 13:33:29 crc kubenswrapper[4815]: I0225 13:33:29.113484 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-92bcq" event={"ID":"801eb953-96a0-4272-9f81-13e80e61f75a","Type":"ContainerStarted","Data":"3b5902325b40f2708b8560a2460a716e504d5b15cd539250de1fb9865c65744c"} Feb 25 13:33:29 crc kubenswrapper[4815]: I0225 13:33:29.113494 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-92bcq" event={"ID":"801eb953-96a0-4272-9f81-13e80e61f75a","Type":"ContainerStarted","Data":"0bd30a89c58816d84831abc917dae2377cfac300cbb08d18dd551f6a2fb467dd"} Feb 25 13:33:29 crc kubenswrapper[4815]: I0225 13:33:29.113522 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-92bcq" event={"ID":"801eb953-96a0-4272-9f81-13e80e61f75a","Type":"ContainerStarted","Data":"aa3d32ac5e2070389015ea2c6565f423b4a807634a0720fd2784cbf3b5650065"} Feb 25 13:33:29 crc kubenswrapper[4815]: I0225 13:33:29.143694 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-92bcq" podStartSLOduration=5.97915019 podStartE2EDuration="12.143674781s" podCreationTimestamp="2026-02-25 13:33:17 +0000 UTC" firstStartedPulling="2026-02-25 13:33:18.762044716 +0000 UTC m=+756.563142770" lastFinishedPulling="2026-02-25 13:33:24.926569287 +0000 UTC m=+762.727667361" observedRunningTime="2026-02-25 13:33:29.141284296 +0000 UTC m=+766.942382390" watchObservedRunningTime="2026-02-25 13:33:29.143674781 +0000 UTC m=+766.944772845" Feb 25 13:33:33 crc kubenswrapper[4815]: I0225 13:33:33.620273 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:33 crc kubenswrapper[4815]: I0225 13:33:33.654460 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:38 crc kubenswrapper[4815]: I0225 13:33:38.035467 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-mcphd" Feb 25 13:33:38 crc kubenswrapper[4815]: I0225 13:33:38.622708 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-92bcq" Feb 25 13:33:38 crc kubenswrapper[4815]: I0225 13:33:38.707378 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-qxvv8" Feb 25 13:33:41 crc kubenswrapper[4815]: I0225 13:33:41.448957 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-k2dbq"] Feb 25 13:33:41 crc kubenswrapper[4815]: I0225 13:33:41.452181 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k2dbq" Feb 25 13:33:41 crc kubenswrapper[4815]: I0225 13:33:41.454748 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 25 13:33:41 crc kubenswrapper[4815]: I0225 13:33:41.455046 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-6msrx" Feb 25 13:33:41 crc kubenswrapper[4815]: I0225 13:33:41.460238 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-k2dbq"] Feb 25 13:33:41 crc kubenswrapper[4815]: I0225 13:33:41.467039 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 25 13:33:41 crc kubenswrapper[4815]: I0225 13:33:41.594111 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bxzz\" (UniqueName: \"kubernetes.io/projected/44c3673b-4dd8-4148-9351-e25a3fda3680-kube-api-access-8bxzz\") pod \"openstack-operator-index-k2dbq\" (UID: \"44c3673b-4dd8-4148-9351-e25a3fda3680\") " pod="openstack-operators/openstack-operator-index-k2dbq" Feb 25 13:33:41 crc kubenswrapper[4815]: I0225 13:33:41.695825 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bxzz\" (UniqueName: \"kubernetes.io/projected/44c3673b-4dd8-4148-9351-e25a3fda3680-kube-api-access-8bxzz\") pod \"openstack-operator-index-k2dbq\" (UID: \"44c3673b-4dd8-4148-9351-e25a3fda3680\") " pod="openstack-operators/openstack-operator-index-k2dbq" Feb 25 13:33:41 crc kubenswrapper[4815]: I0225 13:33:41.714971 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bxzz\" (UniqueName: \"kubernetes.io/projected/44c3673b-4dd8-4148-9351-e25a3fda3680-kube-api-access-8bxzz\") pod \"openstack-operator-index-k2dbq\" (UID: \"44c3673b-4dd8-4148-9351-e25a3fda3680\") " pod="openstack-operators/openstack-operator-index-k2dbq" Feb 25 13:33:41 crc kubenswrapper[4815]: I0225 13:33:41.773265 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k2dbq" Feb 25 13:33:42 crc kubenswrapper[4815]: I0225 13:33:42.217179 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-k2dbq"] Feb 25 13:33:43 crc kubenswrapper[4815]: I0225 13:33:43.212302 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k2dbq" event={"ID":"44c3673b-4dd8-4148-9351-e25a3fda3680","Type":"ContainerStarted","Data":"2f068a1b2d99d3d5dd0bde24a0b03d06f5ea703d51c50ae3e97b570a52eb6528"} Feb 25 13:33:44 crc kubenswrapper[4815]: I0225 13:33:44.821197 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-k2dbq"] Feb 25 13:33:45 crc kubenswrapper[4815]: I0225 13:33:45.229880 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k2dbq" event={"ID":"44c3673b-4dd8-4148-9351-e25a3fda3680","Type":"ContainerStarted","Data":"4b12ee27c67ea7797aed07f91bc19293f46bdea89f5c93c9e835a8eef0a54df6"} Feb 25 13:33:45 crc kubenswrapper[4815]: I0225 13:33:45.419040 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-k2dbq" podStartSLOduration=2.273034939 podStartE2EDuration="4.41900754s" podCreationTimestamp="2026-02-25 13:33:41 +0000 UTC" firstStartedPulling="2026-02-25 13:33:42.232058155 +0000 UTC m=+780.033156199" lastFinishedPulling="2026-02-25 13:33:44.378030746 +0000 UTC m=+782.179128800" observedRunningTime="2026-02-25 13:33:45.2577544 +0000 UTC m=+783.058852504" watchObservedRunningTime="2026-02-25 13:33:45.41900754 +0000 UTC m=+783.220105644" Feb 25 13:33:45 crc kubenswrapper[4815]: I0225 13:33:45.428544 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-fgf9z"] Feb 25 13:33:45 crc kubenswrapper[4815]: I0225 13:33:45.430092 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fgf9z" Feb 25 13:33:45 crc kubenswrapper[4815]: I0225 13:33:45.436001 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fgf9z"] Feb 25 13:33:45 crc kubenswrapper[4815]: I0225 13:33:45.575486 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-776rt\" (UniqueName: \"kubernetes.io/projected/9a2ffc78-6a9a-45b2-81ab-bfe65898a0cd-kube-api-access-776rt\") pod \"openstack-operator-index-fgf9z\" (UID: \"9a2ffc78-6a9a-45b2-81ab-bfe65898a0cd\") " pod="openstack-operators/openstack-operator-index-fgf9z" Feb 25 13:33:45 crc kubenswrapper[4815]: I0225 13:33:45.676698 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-776rt\" (UniqueName: \"kubernetes.io/projected/9a2ffc78-6a9a-45b2-81ab-bfe65898a0cd-kube-api-access-776rt\") pod \"openstack-operator-index-fgf9z\" (UID: \"9a2ffc78-6a9a-45b2-81ab-bfe65898a0cd\") " pod="openstack-operators/openstack-operator-index-fgf9z" Feb 25 13:33:45 crc kubenswrapper[4815]: I0225 13:33:45.701571 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-776rt\" (UniqueName: \"kubernetes.io/projected/9a2ffc78-6a9a-45b2-81ab-bfe65898a0cd-kube-api-access-776rt\") pod \"openstack-operator-index-fgf9z\" (UID: \"9a2ffc78-6a9a-45b2-81ab-bfe65898a0cd\") " pod="openstack-operators/openstack-operator-index-fgf9z" Feb 25 13:33:45 crc kubenswrapper[4815]: I0225 13:33:45.794799 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fgf9z" Feb 25 13:33:46 crc kubenswrapper[4815]: I0225 13:33:46.008805 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fgf9z"] Feb 25 13:33:46 crc kubenswrapper[4815]: W0225 13:33:46.020729 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a2ffc78_6a9a_45b2_81ab_bfe65898a0cd.slice/crio-2bcab0eafa9cf66257db3bfcbb4704e371838b4dd40f35e7c549db9d0d0f75ba WatchSource:0}: Error finding container 2bcab0eafa9cf66257db3bfcbb4704e371838b4dd40f35e7c549db9d0d0f75ba: Status 404 returned error can't find the container with id 2bcab0eafa9cf66257db3bfcbb4704e371838b4dd40f35e7c549db9d0d0f75ba Feb 25 13:33:46 crc kubenswrapper[4815]: I0225 13:33:46.238288 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fgf9z" event={"ID":"9a2ffc78-6a9a-45b2-81ab-bfe65898a0cd","Type":"ContainerStarted","Data":"55c4c2409136e013843c892bbf84bfac922cbdfff8be76d857211cebfe9cc5cd"} Feb 25 13:33:46 crc kubenswrapper[4815]: I0225 13:33:46.238933 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fgf9z" event={"ID":"9a2ffc78-6a9a-45b2-81ab-bfe65898a0cd","Type":"ContainerStarted","Data":"2bcab0eafa9cf66257db3bfcbb4704e371838b4dd40f35e7c549db9d0d0f75ba"} Feb 25 13:33:46 crc kubenswrapper[4815]: I0225 13:33:46.238427 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-k2dbq" podUID="44c3673b-4dd8-4148-9351-e25a3fda3680" containerName="registry-server" containerID="cri-o://4b12ee27c67ea7797aed07f91bc19293f46bdea89f5c93c9e835a8eef0a54df6" gracePeriod=2 Feb 25 13:33:46 crc kubenswrapper[4815]: I0225 13:33:46.265800 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-fgf9z" podStartSLOduration=1.2159455160000001 podStartE2EDuration="1.265743107s" podCreationTimestamp="2026-02-25 13:33:45 +0000 UTC" firstStartedPulling="2026-02-25 13:33:46.024754005 +0000 UTC m=+783.825852059" lastFinishedPulling="2026-02-25 13:33:46.074551566 +0000 UTC m=+783.875649650" observedRunningTime="2026-02-25 13:33:46.255248063 +0000 UTC m=+784.056346147" watchObservedRunningTime="2026-02-25 13:33:46.265743107 +0000 UTC m=+784.066841201" Feb 25 13:33:46 crc kubenswrapper[4815]: I0225 13:33:46.672173 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k2dbq" Feb 25 13:33:46 crc kubenswrapper[4815]: I0225 13:33:46.794706 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bxzz\" (UniqueName: \"kubernetes.io/projected/44c3673b-4dd8-4148-9351-e25a3fda3680-kube-api-access-8bxzz\") pod \"44c3673b-4dd8-4148-9351-e25a3fda3680\" (UID: \"44c3673b-4dd8-4148-9351-e25a3fda3680\") " Feb 25 13:33:46 crc kubenswrapper[4815]: I0225 13:33:46.800926 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44c3673b-4dd8-4148-9351-e25a3fda3680-kube-api-access-8bxzz" (OuterVolumeSpecName: "kube-api-access-8bxzz") pod "44c3673b-4dd8-4148-9351-e25a3fda3680" (UID: "44c3673b-4dd8-4148-9351-e25a3fda3680"). InnerVolumeSpecName "kube-api-access-8bxzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:33:46 crc kubenswrapper[4815]: I0225 13:33:46.896017 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bxzz\" (UniqueName: \"kubernetes.io/projected/44c3673b-4dd8-4148-9351-e25a3fda3680-kube-api-access-8bxzz\") on node \"crc\" DevicePath \"\"" Feb 25 13:33:47 crc kubenswrapper[4815]: I0225 13:33:47.246958 4815 generic.go:334] "Generic (PLEG): container finished" podID="44c3673b-4dd8-4148-9351-e25a3fda3680" containerID="4b12ee27c67ea7797aed07f91bc19293f46bdea89f5c93c9e835a8eef0a54df6" exitCode=0 Feb 25 13:33:47 crc kubenswrapper[4815]: I0225 13:33:47.247051 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k2dbq" Feb 25 13:33:47 crc kubenswrapper[4815]: I0225 13:33:47.247076 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k2dbq" event={"ID":"44c3673b-4dd8-4148-9351-e25a3fda3680","Type":"ContainerDied","Data":"4b12ee27c67ea7797aed07f91bc19293f46bdea89f5c93c9e835a8eef0a54df6"} Feb 25 13:33:47 crc kubenswrapper[4815]: I0225 13:33:47.247872 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k2dbq" event={"ID":"44c3673b-4dd8-4148-9351-e25a3fda3680","Type":"ContainerDied","Data":"2f068a1b2d99d3d5dd0bde24a0b03d06f5ea703d51c50ae3e97b570a52eb6528"} Feb 25 13:33:47 crc kubenswrapper[4815]: I0225 13:33:47.247921 4815 scope.go:117] "RemoveContainer" containerID="4b12ee27c67ea7797aed07f91bc19293f46bdea89f5c93c9e835a8eef0a54df6" Feb 25 13:33:47 crc kubenswrapper[4815]: I0225 13:33:47.276767 4815 scope.go:117] "RemoveContainer" containerID="4b12ee27c67ea7797aed07f91bc19293f46bdea89f5c93c9e835a8eef0a54df6" Feb 25 13:33:47 crc kubenswrapper[4815]: E0225 13:33:47.277870 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b12ee27c67ea7797aed07f91bc19293f46bdea89f5c93c9e835a8eef0a54df6\": container with ID starting with 4b12ee27c67ea7797aed07f91bc19293f46bdea89f5c93c9e835a8eef0a54df6 not found: ID does not exist" containerID="4b12ee27c67ea7797aed07f91bc19293f46bdea89f5c93c9e835a8eef0a54df6" Feb 25 13:33:47 crc kubenswrapper[4815]: I0225 13:33:47.277940 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b12ee27c67ea7797aed07f91bc19293f46bdea89f5c93c9e835a8eef0a54df6"} err="failed to get container status \"4b12ee27c67ea7797aed07f91bc19293f46bdea89f5c93c9e835a8eef0a54df6\": rpc error: code = NotFound desc = could not find container \"4b12ee27c67ea7797aed07f91bc19293f46bdea89f5c93c9e835a8eef0a54df6\": container with ID starting with 4b12ee27c67ea7797aed07f91bc19293f46bdea89f5c93c9e835a8eef0a54df6 not found: ID does not exist" Feb 25 13:33:47 crc kubenswrapper[4815]: I0225 13:33:47.282186 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-k2dbq"] Feb 25 13:33:47 crc kubenswrapper[4815]: I0225 13:33:47.290726 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-k2dbq"] Feb 25 13:33:48 crc kubenswrapper[4815]: I0225 13:33:48.946863 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44c3673b-4dd8-4148-9351-e25a3fda3680" path="/var/lib/kubelet/pods/44c3673b-4dd8-4148-9351-e25a3fda3680/volumes" Feb 25 13:33:51 crc kubenswrapper[4815]: I0225 13:33:51.342433 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:33:51 crc kubenswrapper[4815]: I0225 13:33:51.343040 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:33:55 crc kubenswrapper[4815]: I0225 13:33:55.795711 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-fgf9z" Feb 25 13:33:55 crc kubenswrapper[4815]: I0225 13:33:55.796134 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-fgf9z" Feb 25 13:33:55 crc kubenswrapper[4815]: I0225 13:33:55.838981 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-fgf9z" Feb 25 13:33:56 crc kubenswrapper[4815]: I0225 13:33:56.366965 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-fgf9z" Feb 25 13:34:00 crc kubenswrapper[4815]: I0225 13:34:00.154755 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533774-tm57d"] Feb 25 13:34:00 crc kubenswrapper[4815]: E0225 13:34:00.155748 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44c3673b-4dd8-4148-9351-e25a3fda3680" containerName="registry-server" Feb 25 13:34:00 crc kubenswrapper[4815]: I0225 13:34:00.155781 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="44c3673b-4dd8-4148-9351-e25a3fda3680" containerName="registry-server" Feb 25 13:34:00 crc kubenswrapper[4815]: I0225 13:34:00.156084 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="44c3673b-4dd8-4148-9351-e25a3fda3680" containerName="registry-server" Feb 25 13:34:00 crc kubenswrapper[4815]: I0225 13:34:00.157015 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533774-tm57d" Feb 25 13:34:00 crc kubenswrapper[4815]: I0225 13:34:00.159579 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:34:00 crc kubenswrapper[4815]: I0225 13:34:00.159922 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:34:00 crc kubenswrapper[4815]: I0225 13:34:00.160694 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:34:00 crc kubenswrapper[4815]: I0225 13:34:00.205986 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533774-tm57d"] Feb 25 13:34:00 crc kubenswrapper[4815]: I0225 13:34:00.302955 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rln68\" (UniqueName: \"kubernetes.io/projected/a83fc2e6-4dca-419c-ac69-1eb0f160ba73-kube-api-access-rln68\") pod \"auto-csr-approver-29533774-tm57d\" (UID: \"a83fc2e6-4dca-419c-ac69-1eb0f160ba73\") " pod="openshift-infra/auto-csr-approver-29533774-tm57d" Feb 25 13:34:00 crc kubenswrapper[4815]: I0225 13:34:00.404339 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rln68\" (UniqueName: \"kubernetes.io/projected/a83fc2e6-4dca-419c-ac69-1eb0f160ba73-kube-api-access-rln68\") pod \"auto-csr-approver-29533774-tm57d\" (UID: \"a83fc2e6-4dca-419c-ac69-1eb0f160ba73\") " pod="openshift-infra/auto-csr-approver-29533774-tm57d" Feb 25 13:34:00 crc kubenswrapper[4815]: I0225 13:34:00.430173 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rln68\" (UniqueName: \"kubernetes.io/projected/a83fc2e6-4dca-419c-ac69-1eb0f160ba73-kube-api-access-rln68\") pod \"auto-csr-approver-29533774-tm57d\" (UID: \"a83fc2e6-4dca-419c-ac69-1eb0f160ba73\") " pod="openshift-infra/auto-csr-approver-29533774-tm57d" Feb 25 13:34:00 crc kubenswrapper[4815]: I0225 13:34:00.518043 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533774-tm57d" Feb 25 13:34:00 crc kubenswrapper[4815]: I0225 13:34:00.983557 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533774-tm57d"] Feb 25 13:34:01 crc kubenswrapper[4815]: I0225 13:34:01.378445 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533774-tm57d" event={"ID":"a83fc2e6-4dca-419c-ac69-1eb0f160ba73","Type":"ContainerStarted","Data":"867af7aa72c5452c2b2fe5069a2ed70f542918a3614574ba53a7fb5afc6e2fc7"} Feb 25 13:34:02 crc kubenswrapper[4815]: I0225 13:34:02.387322 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533774-tm57d" event={"ID":"a83fc2e6-4dca-419c-ac69-1eb0f160ba73","Type":"ContainerStarted","Data":"bb4fedbfd8b9b49fa52cde59b50a74925e6f86d35b4e160a2acf3b4a245a3bb5"} Feb 25 13:34:02 crc kubenswrapper[4815]: I0225 13:34:02.411606 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533774-tm57d" podStartSLOduration=1.494052236 podStartE2EDuration="2.411588092s" podCreationTimestamp="2026-02-25 13:34:00 +0000 UTC" firstStartedPulling="2026-02-25 13:34:00.991703066 +0000 UTC m=+798.792801150" lastFinishedPulling="2026-02-25 13:34:01.909238922 +0000 UTC m=+799.710337006" observedRunningTime="2026-02-25 13:34:02.405813589 +0000 UTC m=+800.206911653" watchObservedRunningTime="2026-02-25 13:34:02.411588092 +0000 UTC m=+800.212686146" Feb 25 13:34:03 crc kubenswrapper[4815]: I0225 13:34:03.308988 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d"] Feb 25 13:34:03 crc kubenswrapper[4815]: I0225 13:34:03.313041 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" Feb 25 13:34:03 crc kubenswrapper[4815]: I0225 13:34:03.316891 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-jvd9b" Feb 25 13:34:03 crc kubenswrapper[4815]: I0225 13:34:03.321538 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d"] Feb 25 13:34:03 crc kubenswrapper[4815]: I0225 13:34:03.395738 4815 generic.go:334] "Generic (PLEG): container finished" podID="a83fc2e6-4dca-419c-ac69-1eb0f160ba73" containerID="bb4fedbfd8b9b49fa52cde59b50a74925e6f86d35b4e160a2acf3b4a245a3bb5" exitCode=0 Feb 25 13:34:03 crc kubenswrapper[4815]: I0225 13:34:03.396323 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533774-tm57d" event={"ID":"a83fc2e6-4dca-419c-ac69-1eb0f160ba73","Type":"ContainerDied","Data":"bb4fedbfd8b9b49fa52cde59b50a74925e6f86d35b4e160a2acf3b4a245a3bb5"} Feb 25 13:34:03 crc kubenswrapper[4815]: I0225 13:34:03.453933 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcrv2\" (UniqueName: \"kubernetes.io/projected/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-kube-api-access-dcrv2\") pod \"37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d\" (UID: \"01da6351-8eb8-427f-aaa2-73d2a7bcaac6\") " pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" Feb 25 13:34:03 crc kubenswrapper[4815]: I0225 13:34:03.453982 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-util\") pod \"37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d\" (UID: \"01da6351-8eb8-427f-aaa2-73d2a7bcaac6\") " pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" Feb 25 13:34:03 crc kubenswrapper[4815]: I0225 13:34:03.454015 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-bundle\") pod \"37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d\" (UID: \"01da6351-8eb8-427f-aaa2-73d2a7bcaac6\") " pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" Feb 25 13:34:03 crc kubenswrapper[4815]: I0225 13:34:03.555062 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcrv2\" (UniqueName: \"kubernetes.io/projected/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-kube-api-access-dcrv2\") pod \"37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d\" (UID: \"01da6351-8eb8-427f-aaa2-73d2a7bcaac6\") " pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" Feb 25 13:34:03 crc kubenswrapper[4815]: I0225 13:34:03.555118 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-util\") pod \"37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d\" (UID: \"01da6351-8eb8-427f-aaa2-73d2a7bcaac6\") " pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" Feb 25 13:34:03 crc kubenswrapper[4815]: I0225 13:34:03.555149 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-bundle\") pod \"37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d\" (UID: \"01da6351-8eb8-427f-aaa2-73d2a7bcaac6\") " pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" Feb 25 13:34:03 crc kubenswrapper[4815]: I0225 13:34:03.555675 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-bundle\") pod \"37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d\" (UID: \"01da6351-8eb8-427f-aaa2-73d2a7bcaac6\") " pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" Feb 25 13:34:03 crc kubenswrapper[4815]: I0225 13:34:03.555723 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-util\") pod \"37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d\" (UID: \"01da6351-8eb8-427f-aaa2-73d2a7bcaac6\") " pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" Feb 25 13:34:03 crc kubenswrapper[4815]: I0225 13:34:03.582172 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcrv2\" (UniqueName: \"kubernetes.io/projected/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-kube-api-access-dcrv2\") pod \"37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d\" (UID: \"01da6351-8eb8-427f-aaa2-73d2a7bcaac6\") " pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" Feb 25 13:34:03 crc kubenswrapper[4815]: I0225 13:34:03.643254 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" Feb 25 13:34:04 crc kubenswrapper[4815]: I0225 13:34:04.141548 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d"] Feb 25 13:34:04 crc kubenswrapper[4815]: W0225 13:34:04.147926 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01da6351_8eb8_427f_aaa2_73d2a7bcaac6.slice/crio-5a8be0fabcaac00f055b1c21c8e7f3efe2c2db2de111b29ba9cca3e9f8448fb4 WatchSource:0}: Error finding container 5a8be0fabcaac00f055b1c21c8e7f3efe2c2db2de111b29ba9cca3e9f8448fb4: Status 404 returned error can't find the container with id 5a8be0fabcaac00f055b1c21c8e7f3efe2c2db2de111b29ba9cca3e9f8448fb4 Feb 25 13:34:04 crc kubenswrapper[4815]: I0225 13:34:04.410209 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" event={"ID":"01da6351-8eb8-427f-aaa2-73d2a7bcaac6","Type":"ContainerStarted","Data":"d3b15c8186528cebd668ff852a5851c797107812b9c00a1ab9da5ec19fdb8324"} Feb 25 13:34:04 crc kubenswrapper[4815]: I0225 13:34:04.410740 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" event={"ID":"01da6351-8eb8-427f-aaa2-73d2a7bcaac6","Type":"ContainerStarted","Data":"5a8be0fabcaac00f055b1c21c8e7f3efe2c2db2de111b29ba9cca3e9f8448fb4"} Feb 25 13:34:04 crc kubenswrapper[4815]: I0225 13:34:04.704460 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533774-tm57d" Feb 25 13:34:04 crc kubenswrapper[4815]: I0225 13:34:04.776693 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rln68\" (UniqueName: \"kubernetes.io/projected/a83fc2e6-4dca-419c-ac69-1eb0f160ba73-kube-api-access-rln68\") pod \"a83fc2e6-4dca-419c-ac69-1eb0f160ba73\" (UID: \"a83fc2e6-4dca-419c-ac69-1eb0f160ba73\") " Feb 25 13:34:04 crc kubenswrapper[4815]: I0225 13:34:04.785751 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a83fc2e6-4dca-419c-ac69-1eb0f160ba73-kube-api-access-rln68" (OuterVolumeSpecName: "kube-api-access-rln68") pod "a83fc2e6-4dca-419c-ac69-1eb0f160ba73" (UID: "a83fc2e6-4dca-419c-ac69-1eb0f160ba73"). InnerVolumeSpecName "kube-api-access-rln68". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:34:04 crc kubenswrapper[4815]: I0225 13:34:04.878628 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rln68\" (UniqueName: \"kubernetes.io/projected/a83fc2e6-4dca-419c-ac69-1eb0f160ba73-kube-api-access-rln68\") on node \"crc\" DevicePath \"\"" Feb 25 13:34:05 crc kubenswrapper[4815]: I0225 13:34:05.422978 4815 generic.go:334] "Generic (PLEG): container finished" podID="01da6351-8eb8-427f-aaa2-73d2a7bcaac6" containerID="d3b15c8186528cebd668ff852a5851c797107812b9c00a1ab9da5ec19fdb8324" exitCode=0 Feb 25 13:34:05 crc kubenswrapper[4815]: I0225 13:34:05.423101 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" event={"ID":"01da6351-8eb8-427f-aaa2-73d2a7bcaac6","Type":"ContainerDied","Data":"d3b15c8186528cebd668ff852a5851c797107812b9c00a1ab9da5ec19fdb8324"} Feb 25 13:34:05 crc kubenswrapper[4815]: I0225 13:34:05.426587 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533774-tm57d" event={"ID":"a83fc2e6-4dca-419c-ac69-1eb0f160ba73","Type":"ContainerDied","Data":"867af7aa72c5452c2b2fe5069a2ed70f542918a3614574ba53a7fb5afc6e2fc7"} Feb 25 13:34:05 crc kubenswrapper[4815]: I0225 13:34:05.426713 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="867af7aa72c5452c2b2fe5069a2ed70f542918a3614574ba53a7fb5afc6e2fc7" Feb 25 13:34:05 crc kubenswrapper[4815]: I0225 13:34:05.426822 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533774-tm57d" Feb 25 13:34:05 crc kubenswrapper[4815]: I0225 13:34:05.795603 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533768-b4gbs"] Feb 25 13:34:05 crc kubenswrapper[4815]: I0225 13:34:05.809878 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533768-b4gbs"] Feb 25 13:34:06 crc kubenswrapper[4815]: I0225 13:34:06.436701 4815 generic.go:334] "Generic (PLEG): container finished" podID="01da6351-8eb8-427f-aaa2-73d2a7bcaac6" containerID="b239883a92186ac0a65e58d024453b52a6c1a8b12f066384babc0352565cd569" exitCode=0 Feb 25 13:34:06 crc kubenswrapper[4815]: I0225 13:34:06.436810 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" event={"ID":"01da6351-8eb8-427f-aaa2-73d2a7bcaac6","Type":"ContainerDied","Data":"b239883a92186ac0a65e58d024453b52a6c1a8b12f066384babc0352565cd569"} Feb 25 13:34:06 crc kubenswrapper[4815]: I0225 13:34:06.949475 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02dfab6b-ba3d-4288-b8f2-e41422e7cc78" path="/var/lib/kubelet/pods/02dfab6b-ba3d-4288-b8f2-e41422e7cc78/volumes" Feb 25 13:34:07 crc kubenswrapper[4815]: I0225 13:34:07.449170 4815 generic.go:334] "Generic (PLEG): container finished" podID="01da6351-8eb8-427f-aaa2-73d2a7bcaac6" containerID="5f8deb49b303dc39713415ff3da3938903f7db85615df9327d1c43f7f4aa202f" exitCode=0 Feb 25 13:34:07 crc kubenswrapper[4815]: I0225 13:34:07.449206 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" event={"ID":"01da6351-8eb8-427f-aaa2-73d2a7bcaac6","Type":"ContainerDied","Data":"5f8deb49b303dc39713415ff3da3938903f7db85615df9327d1c43f7f4aa202f"} Feb 25 13:34:08 crc kubenswrapper[4815]: I0225 13:34:08.736821 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" Feb 25 13:34:08 crc kubenswrapper[4815]: I0225 13:34:08.837825 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcrv2\" (UniqueName: \"kubernetes.io/projected/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-kube-api-access-dcrv2\") pod \"01da6351-8eb8-427f-aaa2-73d2a7bcaac6\" (UID: \"01da6351-8eb8-427f-aaa2-73d2a7bcaac6\") " Feb 25 13:34:08 crc kubenswrapper[4815]: I0225 13:34:08.838093 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-util\") pod \"01da6351-8eb8-427f-aaa2-73d2a7bcaac6\" (UID: \"01da6351-8eb8-427f-aaa2-73d2a7bcaac6\") " Feb 25 13:34:08 crc kubenswrapper[4815]: I0225 13:34:08.838153 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-bundle\") pod \"01da6351-8eb8-427f-aaa2-73d2a7bcaac6\" (UID: \"01da6351-8eb8-427f-aaa2-73d2a7bcaac6\") " Feb 25 13:34:08 crc kubenswrapper[4815]: I0225 13:34:08.838865 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-bundle" (OuterVolumeSpecName: "bundle") pod "01da6351-8eb8-427f-aaa2-73d2a7bcaac6" (UID: "01da6351-8eb8-427f-aaa2-73d2a7bcaac6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:34:08 crc kubenswrapper[4815]: I0225 13:34:08.846069 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-kube-api-access-dcrv2" (OuterVolumeSpecName: "kube-api-access-dcrv2") pod "01da6351-8eb8-427f-aaa2-73d2a7bcaac6" (UID: "01da6351-8eb8-427f-aaa2-73d2a7bcaac6"). InnerVolumeSpecName "kube-api-access-dcrv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:34:08 crc kubenswrapper[4815]: I0225 13:34:08.859719 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-util" (OuterVolumeSpecName: "util") pod "01da6351-8eb8-427f-aaa2-73d2a7bcaac6" (UID: "01da6351-8eb8-427f-aaa2-73d2a7bcaac6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:34:08 crc kubenswrapper[4815]: I0225 13:34:08.939378 4815 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-util\") on node \"crc\" DevicePath \"\"" Feb 25 13:34:08 crc kubenswrapper[4815]: I0225 13:34:08.939432 4815 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:34:08 crc kubenswrapper[4815]: I0225 13:34:08.939452 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcrv2\" (UniqueName: \"kubernetes.io/projected/01da6351-8eb8-427f-aaa2-73d2a7bcaac6-kube-api-access-dcrv2\") on node \"crc\" DevicePath \"\"" Feb 25 13:34:09 crc kubenswrapper[4815]: I0225 13:34:09.468101 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" event={"ID":"01da6351-8eb8-427f-aaa2-73d2a7bcaac6","Type":"ContainerDied","Data":"5a8be0fabcaac00f055b1c21c8e7f3efe2c2db2de111b29ba9cca3e9f8448fb4"} Feb 25 13:34:09 crc kubenswrapper[4815]: I0225 13:34:09.468169 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a8be0fabcaac00f055b1c21c8e7f3efe2c2db2de111b29ba9cca3e9f8448fb4" Feb 25 13:34:09 crc kubenswrapper[4815]: I0225 13:34:09.468263 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d" Feb 25 13:34:16 crc kubenswrapper[4815]: I0225 13:34:16.205690 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-c7f69bdc4-8vjkl"] Feb 25 13:34:16 crc kubenswrapper[4815]: E0225 13:34:16.206606 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01da6351-8eb8-427f-aaa2-73d2a7bcaac6" containerName="pull" Feb 25 13:34:16 crc kubenswrapper[4815]: I0225 13:34:16.206648 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="01da6351-8eb8-427f-aaa2-73d2a7bcaac6" containerName="pull" Feb 25 13:34:16 crc kubenswrapper[4815]: E0225 13:34:16.206660 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01da6351-8eb8-427f-aaa2-73d2a7bcaac6" containerName="util" Feb 25 13:34:16 crc kubenswrapper[4815]: I0225 13:34:16.206673 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="01da6351-8eb8-427f-aaa2-73d2a7bcaac6" containerName="util" Feb 25 13:34:16 crc kubenswrapper[4815]: E0225 13:34:16.206688 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01da6351-8eb8-427f-aaa2-73d2a7bcaac6" containerName="extract" Feb 25 13:34:16 crc kubenswrapper[4815]: I0225 13:34:16.206700 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="01da6351-8eb8-427f-aaa2-73d2a7bcaac6" containerName="extract" Feb 25 13:34:16 crc kubenswrapper[4815]: E0225 13:34:16.206726 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a83fc2e6-4dca-419c-ac69-1eb0f160ba73" containerName="oc" Feb 25 13:34:16 crc kubenswrapper[4815]: I0225 13:34:16.206739 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a83fc2e6-4dca-419c-ac69-1eb0f160ba73" containerName="oc" Feb 25 13:34:16 crc kubenswrapper[4815]: I0225 13:34:16.206940 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="a83fc2e6-4dca-419c-ac69-1eb0f160ba73" containerName="oc" Feb 25 13:34:16 crc kubenswrapper[4815]: I0225 13:34:16.206964 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="01da6351-8eb8-427f-aaa2-73d2a7bcaac6" containerName="extract" Feb 25 13:34:16 crc kubenswrapper[4815]: I0225 13:34:16.207631 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-c7f69bdc4-8vjkl" Feb 25 13:34:16 crc kubenswrapper[4815]: I0225 13:34:16.209283 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-xcnf6" Feb 25 13:34:16 crc kubenswrapper[4815]: I0225 13:34:16.238968 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-c7f69bdc4-8vjkl"] Feb 25 13:34:16 crc kubenswrapper[4815]: I0225 13:34:16.257681 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkgz8\" (UniqueName: \"kubernetes.io/projected/a1251cc6-4b50-4551-b051-028595a75965-kube-api-access-pkgz8\") pod \"openstack-operator-controller-init-c7f69bdc4-8vjkl\" (UID: \"a1251cc6-4b50-4551-b051-028595a75965\") " pod="openstack-operators/openstack-operator-controller-init-c7f69bdc4-8vjkl" Feb 25 13:34:16 crc kubenswrapper[4815]: I0225 13:34:16.358408 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkgz8\" (UniqueName: \"kubernetes.io/projected/a1251cc6-4b50-4551-b051-028595a75965-kube-api-access-pkgz8\") pod \"openstack-operator-controller-init-c7f69bdc4-8vjkl\" (UID: \"a1251cc6-4b50-4551-b051-028595a75965\") " pod="openstack-operators/openstack-operator-controller-init-c7f69bdc4-8vjkl" Feb 25 13:34:16 crc kubenswrapper[4815]: I0225 13:34:16.377543 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkgz8\" (UniqueName: \"kubernetes.io/projected/a1251cc6-4b50-4551-b051-028595a75965-kube-api-access-pkgz8\") pod \"openstack-operator-controller-init-c7f69bdc4-8vjkl\" (UID: \"a1251cc6-4b50-4551-b051-028595a75965\") " pod="openstack-operators/openstack-operator-controller-init-c7f69bdc4-8vjkl" Feb 25 13:34:16 crc kubenswrapper[4815]: I0225 13:34:16.526157 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-c7f69bdc4-8vjkl" Feb 25 13:34:16 crc kubenswrapper[4815]: I0225 13:34:16.778488 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-c7f69bdc4-8vjkl"] Feb 25 13:34:17 crc kubenswrapper[4815]: I0225 13:34:17.522157 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-c7f69bdc4-8vjkl" event={"ID":"a1251cc6-4b50-4551-b051-028595a75965","Type":"ContainerStarted","Data":"357e7956312ac767459beef88ba9d80cbf07797ac8a2dd28402ea7e77c9f52a7"} Feb 25 13:34:21 crc kubenswrapper[4815]: I0225 13:34:21.342178 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:34:21 crc kubenswrapper[4815]: I0225 13:34:21.342666 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:34:21 crc kubenswrapper[4815]: I0225 13:34:21.342705 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:34:21 crc kubenswrapper[4815]: I0225 13:34:21.343221 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6557b63a0f423077dc6748b76d4eac34119edc35e845268afc85f7e306b0281a"} pod="openshift-machine-config-operator/machine-config-daemon-frc75" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 13:34:21 crc kubenswrapper[4815]: I0225 13:34:21.343273 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" containerID="cri-o://6557b63a0f423077dc6748b76d4eac34119edc35e845268afc85f7e306b0281a" gracePeriod=600 Feb 25 13:34:21 crc kubenswrapper[4815]: I0225 13:34:21.551773 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-c7f69bdc4-8vjkl" event={"ID":"a1251cc6-4b50-4551-b051-028595a75965","Type":"ContainerStarted","Data":"a752f85e433084ce835a46119ea4ed49e841da08072f60c32d85ffb6efd0a89a"} Feb 25 13:34:21 crc kubenswrapper[4815]: I0225 13:34:21.552068 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-c7f69bdc4-8vjkl" Feb 25 13:34:21 crc kubenswrapper[4815]: I0225 13:34:21.556592 4815 generic.go:334] "Generic (PLEG): container finished" podID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerID="6557b63a0f423077dc6748b76d4eac34119edc35e845268afc85f7e306b0281a" exitCode=0 Feb 25 13:34:21 crc kubenswrapper[4815]: I0225 13:34:21.556644 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerDied","Data":"6557b63a0f423077dc6748b76d4eac34119edc35e845268afc85f7e306b0281a"} Feb 25 13:34:21 crc kubenswrapper[4815]: I0225 13:34:21.556811 4815 scope.go:117] "RemoveContainer" containerID="4e00e443cf7572407e0857913be27d833d029fb2c70b14c349ec691336c7312d" Feb 25 13:34:21 crc kubenswrapper[4815]: I0225 13:34:21.594412 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-c7f69bdc4-8vjkl" podStartSLOduration=1.9060654910000001 podStartE2EDuration="5.594388366s" podCreationTimestamp="2026-02-25 13:34:16 +0000 UTC" firstStartedPulling="2026-02-25 13:34:16.787316278 +0000 UTC m=+814.588414332" lastFinishedPulling="2026-02-25 13:34:20.475639153 +0000 UTC m=+818.276737207" observedRunningTime="2026-02-25 13:34:21.594166729 +0000 UTC m=+819.395264873" watchObservedRunningTime="2026-02-25 13:34:21.594388366 +0000 UTC m=+819.395486460" Feb 25 13:34:22 crc kubenswrapper[4815]: I0225 13:34:22.568976 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerStarted","Data":"65cc59f637c0af5e4fc8efb7df93ca963c3aa3c1788d5699757ce0a0287dce17"} Feb 25 13:34:26 crc kubenswrapper[4815]: I0225 13:34:26.529322 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-c7f69bdc4-8vjkl" Feb 25 13:34:43 crc kubenswrapper[4815]: I0225 13:34:43.523095 4815 scope.go:117] "RemoveContainer" containerID="d436db49705ce47a5ea2e0f27945c303583256e6b301f477c0ce993911b28b73" Feb 25 13:34:50 crc kubenswrapper[4815]: I0225 13:34:50.609979 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t9pdn"] Feb 25 13:34:50 crc kubenswrapper[4815]: I0225 13:34:50.611462 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t9pdn" Feb 25 13:34:50 crc kubenswrapper[4815]: I0225 13:34:50.640628 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t9pdn"] Feb 25 13:34:50 crc kubenswrapper[4815]: I0225 13:34:50.658912 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c43e580-fd0a-455b-92e9-c87d13dccccc-catalog-content\") pod \"redhat-marketplace-t9pdn\" (UID: \"7c43e580-fd0a-455b-92e9-c87d13dccccc\") " pod="openshift-marketplace/redhat-marketplace-t9pdn" Feb 25 13:34:50 crc kubenswrapper[4815]: I0225 13:34:50.658964 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgsp2\" (UniqueName: \"kubernetes.io/projected/7c43e580-fd0a-455b-92e9-c87d13dccccc-kube-api-access-mgsp2\") pod \"redhat-marketplace-t9pdn\" (UID: \"7c43e580-fd0a-455b-92e9-c87d13dccccc\") " pod="openshift-marketplace/redhat-marketplace-t9pdn" Feb 25 13:34:50 crc kubenswrapper[4815]: I0225 13:34:50.659039 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c43e580-fd0a-455b-92e9-c87d13dccccc-utilities\") pod \"redhat-marketplace-t9pdn\" (UID: \"7c43e580-fd0a-455b-92e9-c87d13dccccc\") " pod="openshift-marketplace/redhat-marketplace-t9pdn" Feb 25 13:34:50 crc kubenswrapper[4815]: I0225 13:34:50.759935 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c43e580-fd0a-455b-92e9-c87d13dccccc-utilities\") pod \"redhat-marketplace-t9pdn\" (UID: \"7c43e580-fd0a-455b-92e9-c87d13dccccc\") " pod="openshift-marketplace/redhat-marketplace-t9pdn" Feb 25 13:34:50 crc kubenswrapper[4815]: I0225 13:34:50.760013 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c43e580-fd0a-455b-92e9-c87d13dccccc-catalog-content\") pod \"redhat-marketplace-t9pdn\" (UID: \"7c43e580-fd0a-455b-92e9-c87d13dccccc\") " pod="openshift-marketplace/redhat-marketplace-t9pdn" Feb 25 13:34:50 crc kubenswrapper[4815]: I0225 13:34:50.760041 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgsp2\" (UniqueName: \"kubernetes.io/projected/7c43e580-fd0a-455b-92e9-c87d13dccccc-kube-api-access-mgsp2\") pod \"redhat-marketplace-t9pdn\" (UID: \"7c43e580-fd0a-455b-92e9-c87d13dccccc\") " pod="openshift-marketplace/redhat-marketplace-t9pdn" Feb 25 13:34:50 crc kubenswrapper[4815]: I0225 13:34:50.760450 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c43e580-fd0a-455b-92e9-c87d13dccccc-utilities\") pod \"redhat-marketplace-t9pdn\" (UID: \"7c43e580-fd0a-455b-92e9-c87d13dccccc\") " pod="openshift-marketplace/redhat-marketplace-t9pdn" Feb 25 13:34:50 crc kubenswrapper[4815]: I0225 13:34:50.760524 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c43e580-fd0a-455b-92e9-c87d13dccccc-catalog-content\") pod \"redhat-marketplace-t9pdn\" (UID: \"7c43e580-fd0a-455b-92e9-c87d13dccccc\") " pod="openshift-marketplace/redhat-marketplace-t9pdn" Feb 25 13:34:50 crc kubenswrapper[4815]: I0225 13:34:50.792856 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgsp2\" (UniqueName: \"kubernetes.io/projected/7c43e580-fd0a-455b-92e9-c87d13dccccc-kube-api-access-mgsp2\") pod \"redhat-marketplace-t9pdn\" (UID: \"7c43e580-fd0a-455b-92e9-c87d13dccccc\") " pod="openshift-marketplace/redhat-marketplace-t9pdn" Feb 25 13:34:50 crc kubenswrapper[4815]: I0225 13:34:50.929192 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t9pdn" Feb 25 13:34:51 crc kubenswrapper[4815]: I0225 13:34:51.127386 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t9pdn"] Feb 25 13:34:51 crc kubenswrapper[4815]: W0225 13:34:51.134602 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c43e580_fd0a_455b_92e9_c87d13dccccc.slice/crio-3c016f8ce591eb0dbe32a117b723b8e489c48b62b5754e122e2ddbb83d004bf6 WatchSource:0}: Error finding container 3c016f8ce591eb0dbe32a117b723b8e489c48b62b5754e122e2ddbb83d004bf6: Status 404 returned error can't find the container with id 3c016f8ce591eb0dbe32a117b723b8e489c48b62b5754e122e2ddbb83d004bf6 Feb 25 13:34:51 crc kubenswrapper[4815]: I0225 13:34:51.810143 4815 generic.go:334] "Generic (PLEG): container finished" podID="7c43e580-fd0a-455b-92e9-c87d13dccccc" containerID="46850230513cda51c82c76d9278dd5a7c6b196cf196f46d01bc8adc6b10eb94c" exitCode=0 Feb 25 13:34:51 crc kubenswrapper[4815]: I0225 13:34:51.810193 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t9pdn" event={"ID":"7c43e580-fd0a-455b-92e9-c87d13dccccc","Type":"ContainerDied","Data":"46850230513cda51c82c76d9278dd5a7c6b196cf196f46d01bc8adc6b10eb94c"} Feb 25 13:34:51 crc kubenswrapper[4815]: I0225 13:34:51.810222 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t9pdn" event={"ID":"7c43e580-fd0a-455b-92e9-c87d13dccccc","Type":"ContainerStarted","Data":"3c016f8ce591eb0dbe32a117b723b8e489c48b62b5754e122e2ddbb83d004bf6"} Feb 25 13:34:52 crc kubenswrapper[4815]: I0225 13:34:52.818888 4815 generic.go:334] "Generic (PLEG): container finished" podID="7c43e580-fd0a-455b-92e9-c87d13dccccc" containerID="5ec8c649f5cd3841cf2919eaf8fe6f595bcb04e2a6bee85c5e451d2dce16d6af" exitCode=0 Feb 25 13:34:52 crc kubenswrapper[4815]: I0225 13:34:52.818966 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t9pdn" event={"ID":"7c43e580-fd0a-455b-92e9-c87d13dccccc","Type":"ContainerDied","Data":"5ec8c649f5cd3841cf2919eaf8fe6f595bcb04e2a6bee85c5e451d2dce16d6af"} Feb 25 13:34:53 crc kubenswrapper[4815]: I0225 13:34:53.827009 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t9pdn" event={"ID":"7c43e580-fd0a-455b-92e9-c87d13dccccc","Type":"ContainerStarted","Data":"477af7a0fea6b9b41dee24877149c9d883dfb2982c91507b6d577f142e3ad967"} Feb 25 13:34:53 crc kubenswrapper[4815]: I0225 13:34:53.858463 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t9pdn" podStartSLOduration=2.474944448 podStartE2EDuration="3.858445798s" podCreationTimestamp="2026-02-25 13:34:50 +0000 UTC" firstStartedPulling="2026-02-25 13:34:51.812309737 +0000 UTC m=+849.613407801" lastFinishedPulling="2026-02-25 13:34:53.195811057 +0000 UTC m=+850.996909151" observedRunningTime="2026-02-25 13:34:53.854747253 +0000 UTC m=+851.655845357" watchObservedRunningTime="2026-02-25 13:34:53.858445798 +0000 UTC m=+851.659543862" Feb 25 13:34:56 crc kubenswrapper[4815]: I0225 13:34:56.957171 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vt458"] Feb 25 13:34:56 crc kubenswrapper[4815]: I0225 13:34:56.959647 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vt458" Feb 25 13:34:56 crc kubenswrapper[4815]: I0225 13:34:56.987815 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vt458"] Feb 25 13:34:57 crc kubenswrapper[4815]: I0225 13:34:57.056195 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-catalog-content\") pod \"redhat-operators-vt458\" (UID: \"2ff19dbc-fec6-46e8-b36a-e49dde7db51c\") " pod="openshift-marketplace/redhat-operators-vt458" Feb 25 13:34:57 crc kubenswrapper[4815]: I0225 13:34:57.056256 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-utilities\") pod \"redhat-operators-vt458\" (UID: \"2ff19dbc-fec6-46e8-b36a-e49dde7db51c\") " pod="openshift-marketplace/redhat-operators-vt458" Feb 25 13:34:57 crc kubenswrapper[4815]: I0225 13:34:57.056282 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2tjh\" (UniqueName: \"kubernetes.io/projected/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-kube-api-access-k2tjh\") pod \"redhat-operators-vt458\" (UID: \"2ff19dbc-fec6-46e8-b36a-e49dde7db51c\") " pod="openshift-marketplace/redhat-operators-vt458" Feb 25 13:34:57 crc kubenswrapper[4815]: I0225 13:34:57.157673 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-catalog-content\") pod \"redhat-operators-vt458\" (UID: \"2ff19dbc-fec6-46e8-b36a-e49dde7db51c\") " pod="openshift-marketplace/redhat-operators-vt458" Feb 25 13:34:57 crc kubenswrapper[4815]: I0225 13:34:57.157745 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-utilities\") pod \"redhat-operators-vt458\" (UID: \"2ff19dbc-fec6-46e8-b36a-e49dde7db51c\") " pod="openshift-marketplace/redhat-operators-vt458" Feb 25 13:34:57 crc kubenswrapper[4815]: I0225 13:34:57.157777 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2tjh\" (UniqueName: \"kubernetes.io/projected/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-kube-api-access-k2tjh\") pod \"redhat-operators-vt458\" (UID: \"2ff19dbc-fec6-46e8-b36a-e49dde7db51c\") " pod="openshift-marketplace/redhat-operators-vt458" Feb 25 13:34:57 crc kubenswrapper[4815]: I0225 13:34:57.158169 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-catalog-content\") pod \"redhat-operators-vt458\" (UID: \"2ff19dbc-fec6-46e8-b36a-e49dde7db51c\") " pod="openshift-marketplace/redhat-operators-vt458" Feb 25 13:34:57 crc kubenswrapper[4815]: I0225 13:34:57.158319 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-utilities\") pod \"redhat-operators-vt458\" (UID: \"2ff19dbc-fec6-46e8-b36a-e49dde7db51c\") " pod="openshift-marketplace/redhat-operators-vt458" Feb 25 13:34:57 crc kubenswrapper[4815]: I0225 13:34:57.177237 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2tjh\" (UniqueName: \"kubernetes.io/projected/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-kube-api-access-k2tjh\") pod \"redhat-operators-vt458\" (UID: \"2ff19dbc-fec6-46e8-b36a-e49dde7db51c\") " pod="openshift-marketplace/redhat-operators-vt458" Feb 25 13:34:57 crc kubenswrapper[4815]: I0225 13:34:57.282083 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vt458" Feb 25 13:34:57 crc kubenswrapper[4815]: I0225 13:34:57.546607 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vt458"] Feb 25 13:34:57 crc kubenswrapper[4815]: W0225 13:34:57.551840 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ff19dbc_fec6_46e8_b36a_e49dde7db51c.slice/crio-2c528fb19f9c716e34dec868c18069058aabce3c8644457b55ee4a8aeef8ee77 WatchSource:0}: Error finding container 2c528fb19f9c716e34dec868c18069058aabce3c8644457b55ee4a8aeef8ee77: Status 404 returned error can't find the container with id 2c528fb19f9c716e34dec868c18069058aabce3c8644457b55ee4a8aeef8ee77 Feb 25 13:34:57 crc kubenswrapper[4815]: I0225 13:34:57.858269 4815 generic.go:334] "Generic (PLEG): container finished" podID="2ff19dbc-fec6-46e8-b36a-e49dde7db51c" containerID="bfb0aed0c74c3a59dab1f49deec4a2b6ea1beea423b2d384678ea0ce6ccbe7ab" exitCode=0 Feb 25 13:34:57 crc kubenswrapper[4815]: I0225 13:34:57.858476 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vt458" event={"ID":"2ff19dbc-fec6-46e8-b36a-e49dde7db51c","Type":"ContainerDied","Data":"bfb0aed0c74c3a59dab1f49deec4a2b6ea1beea423b2d384678ea0ce6ccbe7ab"} Feb 25 13:34:57 crc kubenswrapper[4815]: I0225 13:34:57.858625 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vt458" event={"ID":"2ff19dbc-fec6-46e8-b36a-e49dde7db51c","Type":"ContainerStarted","Data":"2c528fb19f9c716e34dec868c18069058aabce3c8644457b55ee4a8aeef8ee77"} Feb 25 13:34:59 crc kubenswrapper[4815]: I0225 13:34:59.871469 4815 generic.go:334] "Generic (PLEG): container finished" podID="2ff19dbc-fec6-46e8-b36a-e49dde7db51c" containerID="dc4436fb78751730fc98b1343e2e2fee687e3dc621d385adbb433dfc094401f4" exitCode=0 Feb 25 13:34:59 crc kubenswrapper[4815]: I0225 13:34:59.871594 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vt458" event={"ID":"2ff19dbc-fec6-46e8-b36a-e49dde7db51c","Type":"ContainerDied","Data":"dc4436fb78751730fc98b1343e2e2fee687e3dc621d385adbb433dfc094401f4"} Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.359207 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tddkm"] Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.360869 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tddkm" Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.377458 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tddkm"] Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.501066 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgwvt\" (UniqueName: \"kubernetes.io/projected/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-kube-api-access-cgwvt\") pod \"certified-operators-tddkm\" (UID: \"59dc6b8b-e3e2-4f25-91bf-310d17bbd603\") " pod="openshift-marketplace/certified-operators-tddkm" Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.501450 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-utilities\") pod \"certified-operators-tddkm\" (UID: \"59dc6b8b-e3e2-4f25-91bf-310d17bbd603\") " pod="openshift-marketplace/certified-operators-tddkm" Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.501478 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-catalog-content\") pod \"certified-operators-tddkm\" (UID: \"59dc6b8b-e3e2-4f25-91bf-310d17bbd603\") " pod="openshift-marketplace/certified-operators-tddkm" Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.602526 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgwvt\" (UniqueName: \"kubernetes.io/projected/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-kube-api-access-cgwvt\") pod \"certified-operators-tddkm\" (UID: \"59dc6b8b-e3e2-4f25-91bf-310d17bbd603\") " pod="openshift-marketplace/certified-operators-tddkm" Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.602610 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-utilities\") pod \"certified-operators-tddkm\" (UID: \"59dc6b8b-e3e2-4f25-91bf-310d17bbd603\") " pod="openshift-marketplace/certified-operators-tddkm" Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.602638 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-catalog-content\") pod \"certified-operators-tddkm\" (UID: \"59dc6b8b-e3e2-4f25-91bf-310d17bbd603\") " pod="openshift-marketplace/certified-operators-tddkm" Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.603095 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-utilities\") pod \"certified-operators-tddkm\" (UID: \"59dc6b8b-e3e2-4f25-91bf-310d17bbd603\") " pod="openshift-marketplace/certified-operators-tddkm" Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.603176 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-catalog-content\") pod \"certified-operators-tddkm\" (UID: \"59dc6b8b-e3e2-4f25-91bf-310d17bbd603\") " pod="openshift-marketplace/certified-operators-tddkm" Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.633786 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgwvt\" (UniqueName: \"kubernetes.io/projected/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-kube-api-access-cgwvt\") pod \"certified-operators-tddkm\" (UID: \"59dc6b8b-e3e2-4f25-91bf-310d17bbd603\") " pod="openshift-marketplace/certified-operators-tddkm" Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.685064 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tddkm" Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.902196 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vt458" event={"ID":"2ff19dbc-fec6-46e8-b36a-e49dde7db51c","Type":"ContainerStarted","Data":"949c8fc460788e63ce64de8cdc808e03544a91c144fe4133a9f376e186ac5332"} Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.931332 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t9pdn" Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.931376 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t9pdn" Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.983772 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vt458" podStartSLOduration=2.4838436440000002 podStartE2EDuration="4.983757209s" podCreationTimestamp="2026-02-25 13:34:56 +0000 UTC" firstStartedPulling="2026-02-25 13:34:57.860412396 +0000 UTC m=+855.661510450" lastFinishedPulling="2026-02-25 13:35:00.360325951 +0000 UTC m=+858.161424015" observedRunningTime="2026-02-25 13:35:00.927000145 +0000 UTC m=+858.728098229" watchObservedRunningTime="2026-02-25 13:35:00.983757209 +0000 UTC m=+858.784855263" Feb 25 13:35:00 crc kubenswrapper[4815]: I0225 13:35:00.987606 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tddkm"] Feb 25 13:35:00 crc kubenswrapper[4815]: W0225 13:35:00.993198 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59dc6b8b_e3e2_4f25_91bf_310d17bbd603.slice/crio-9c58fd8b301072c4cf0acf81b294316b94bbd34bd73e480cbe757be9387ab4e3 WatchSource:0}: Error finding container 9c58fd8b301072c4cf0acf81b294316b94bbd34bd73e480cbe757be9387ab4e3: Status 404 returned error can't find the container with id 9c58fd8b301072c4cf0acf81b294316b94bbd34bd73e480cbe757be9387ab4e3 Feb 25 13:35:01 crc kubenswrapper[4815]: I0225 13:35:01.007449 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t9pdn" Feb 25 13:35:01 crc kubenswrapper[4815]: I0225 13:35:01.910412 4815 generic.go:334] "Generic (PLEG): container finished" podID="59dc6b8b-e3e2-4f25-91bf-310d17bbd603" containerID="91ae3ac326d06086bce047b187279b1041e096036daa45a559cb9c58cb9415e9" exitCode=0 Feb 25 13:35:01 crc kubenswrapper[4815]: I0225 13:35:01.910542 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tddkm" event={"ID":"59dc6b8b-e3e2-4f25-91bf-310d17bbd603","Type":"ContainerDied","Data":"91ae3ac326d06086bce047b187279b1041e096036daa45a559cb9c58cb9415e9"} Feb 25 13:35:01 crc kubenswrapper[4815]: I0225 13:35:01.910840 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tddkm" event={"ID":"59dc6b8b-e3e2-4f25-91bf-310d17bbd603","Type":"ContainerStarted","Data":"9c58fd8b301072c4cf0acf81b294316b94bbd34bd73e480cbe757be9387ab4e3"} Feb 25 13:35:01 crc kubenswrapper[4815]: I0225 13:35:01.970464 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t9pdn" Feb 25 13:35:02 crc kubenswrapper[4815]: I0225 13:35:02.918127 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tddkm" event={"ID":"59dc6b8b-e3e2-4f25-91bf-310d17bbd603","Type":"ContainerStarted","Data":"e687322c626dafb096300a9c8ba7e52208ef18d15312da9bcd8afd9d613c2ff3"} Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.179755 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-zwknl"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.180700 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-zwknl" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.182757 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-fctbd" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.184813 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-h9n4d"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.185556 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-h9n4d" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.187536 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-8bl87" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.195347 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-zwknl"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.205040 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-h9n4d"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.212918 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-qm5sw"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.213715 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-qm5sw" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.216058 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-mwg2d" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.221689 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-784b5bb6c5-gw8z5"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.222848 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-gw8z5" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.226102 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-qm5sw"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.229292 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-82h6z" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.238711 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4xns\" (UniqueName: \"kubernetes.io/projected/bace17a8-ce7d-4df9-8259-68931eeecde8-kube-api-access-q4xns\") pod \"cinder-operator-controller-manager-55d77d7b5c-h9n4d\" (UID: \"bace17a8-ce7d-4df9-8259-68931eeecde8\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-h9n4d" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.238784 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqbbs\" (UniqueName: \"kubernetes.io/projected/62c00caf-05cf-4229-9349-341a2b597b12-kube-api-access-mqbbs\") pod \"barbican-operator-controller-manager-868647ff47-zwknl\" (UID: \"62c00caf-05cf-4229-9349-341a2b597b12\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-zwknl" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.252550 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-784b5bb6c5-gw8z5"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.265579 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-csfk2"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.269973 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-csfk2" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.273048 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-rmlsb" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.294869 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-ccl5w"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.296436 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-ccl5w" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.303174 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-7ngxq" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.334356 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-csfk2"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.339822 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp7zm\" (UniqueName: \"kubernetes.io/projected/9fa8caf9-fab3-44d2-9841-593558e9a690-kube-api-access-cp7zm\") pod \"glance-operator-controller-manager-784b5bb6c5-gw8z5\" (UID: \"9fa8caf9-fab3-44d2-9841-593558e9a690\") " pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-gw8z5" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.339869 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqbbs\" (UniqueName: \"kubernetes.io/projected/62c00caf-05cf-4229-9349-341a2b597b12-kube-api-access-mqbbs\") pod \"barbican-operator-controller-manager-868647ff47-zwknl\" (UID: \"62c00caf-05cf-4229-9349-341a2b597b12\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-zwknl" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.339904 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mjgl\" (UniqueName: \"kubernetes.io/projected/50963101-0a9e-453d-925b-ac6f08c545c4-kube-api-access-4mjgl\") pod \"heat-operator-controller-manager-69f49c598c-csfk2\" (UID: \"50963101-0a9e-453d-925b-ac6f08c545c4\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-csfk2" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.339961 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmwmq\" (UniqueName: \"kubernetes.io/projected/c4956173-674d-4b10-9a3f-9996f9190299-kube-api-access-tmwmq\") pod \"designate-operator-controller-manager-6d8bf5c495-qm5sw\" (UID: \"c4956173-674d-4b10-9a3f-9996f9190299\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-qm5sw" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.340011 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grv98\" (UniqueName: \"kubernetes.io/projected/33e59af6-b9bb-4b55-b6a9-c6963e269b94-kube-api-access-grv98\") pod \"horizon-operator-controller-manager-5b9b8895d5-ccl5w\" (UID: \"33e59af6-b9bb-4b55-b6a9-c6963e269b94\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-ccl5w" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.340034 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4xns\" (UniqueName: \"kubernetes.io/projected/bace17a8-ce7d-4df9-8259-68931eeecde8-kube-api-access-q4xns\") pod \"cinder-operator-controller-manager-55d77d7b5c-h9n4d\" (UID: \"bace17a8-ce7d-4df9-8259-68931eeecde8\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-h9n4d" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.358215 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-ccl5w"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.364659 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.365723 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.374953 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.375093 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.375175 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-wg8sf" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.376298 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqbbs\" (UniqueName: \"kubernetes.io/projected/62c00caf-05cf-4229-9349-341a2b597b12-kube-api-access-mqbbs\") pod \"barbican-operator-controller-manager-868647ff47-zwknl\" (UID: \"62c00caf-05cf-4229-9349-341a2b597b12\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-zwknl" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.386037 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-52psd"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.386847 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-52psd" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.390366 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-ns227" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.391369 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4xns\" (UniqueName: \"kubernetes.io/projected/bace17a8-ce7d-4df9-8259-68931eeecde8-kube-api-access-q4xns\") pod \"cinder-operator-controller-manager-55d77d7b5c-h9n4d\" (UID: \"bace17a8-ce7d-4df9-8259-68931eeecde8\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-h9n4d" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.407407 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-52psd"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.419585 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-rx4fg"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.426593 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-rx4fg" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.428117 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-w8jt5" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.435217 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-5fq96"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.436136 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-67d996989d-5fq96" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.440742 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s79s\" (UniqueName: \"kubernetes.io/projected/12b1d896-ab7e-4097-afa4-791d39e1924e-kube-api-access-8s79s\") pod \"ironic-operator-controller-manager-554564d7fc-52psd\" (UID: \"12b1d896-ab7e-4097-afa4-791d39e1924e\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-52psd" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.440788 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp7zm\" (UniqueName: \"kubernetes.io/projected/9fa8caf9-fab3-44d2-9841-593558e9a690-kube-api-access-cp7zm\") pod \"glance-operator-controller-manager-784b5bb6c5-gw8z5\" (UID: \"9fa8caf9-fab3-44d2-9841-593558e9a690\") " pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-gw8z5" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.440843 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mjgl\" (UniqueName: \"kubernetes.io/projected/50963101-0a9e-453d-925b-ac6f08c545c4-kube-api-access-4mjgl\") pod \"heat-operator-controller-manager-69f49c598c-csfk2\" (UID: \"50963101-0a9e-453d-925b-ac6f08c545c4\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-csfk2" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.440868 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert\") pod \"infra-operator-controller-manager-79d975b745-s5ztn\" (UID: \"9a554278-1139-4942-904a-ffc60adde6de\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.440897 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztghv\" (UniqueName: \"kubernetes.io/projected/9a554278-1139-4942-904a-ffc60adde6de-kube-api-access-ztghv\") pod \"infra-operator-controller-manager-79d975b745-s5ztn\" (UID: \"9a554278-1139-4942-904a-ffc60adde6de\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.440944 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmwmq\" (UniqueName: \"kubernetes.io/projected/c4956173-674d-4b10-9a3f-9996f9190299-kube-api-access-tmwmq\") pod \"designate-operator-controller-manager-6d8bf5c495-qm5sw\" (UID: \"c4956173-674d-4b10-9a3f-9996f9190299\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-qm5sw" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.440964 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grv98\" (UniqueName: \"kubernetes.io/projected/33e59af6-b9bb-4b55-b6a9-c6963e269b94-kube-api-access-grv98\") pod \"horizon-operator-controller-manager-5b9b8895d5-ccl5w\" (UID: \"33e59af6-b9bb-4b55-b6a9-c6963e269b94\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-ccl5w" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.446386 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-jxrlh" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.463640 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-5fq96"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.468252 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-rx4fg"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.478624 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6bd4687957-vnb57"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.479440 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-vnb57" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.482202 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grv98\" (UniqueName: \"kubernetes.io/projected/33e59af6-b9bb-4b55-b6a9-c6963e269b94-kube-api-access-grv98\") pod \"horizon-operator-controller-manager-5b9b8895d5-ccl5w\" (UID: \"33e59af6-b9bb-4b55-b6a9-c6963e269b94\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-ccl5w" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.483876 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-wjvvc" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.486562 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmwmq\" (UniqueName: \"kubernetes.io/projected/c4956173-674d-4b10-9a3f-9996f9190299-kube-api-access-tmwmq\") pod \"designate-operator-controller-manager-6d8bf5c495-qm5sw\" (UID: \"c4956173-674d-4b10-9a3f-9996f9190299\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-qm5sw" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.486695 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp7zm\" (UniqueName: \"kubernetes.io/projected/9fa8caf9-fab3-44d2-9841-593558e9a690-kube-api-access-cp7zm\") pod \"glance-operator-controller-manager-784b5bb6c5-gw8z5\" (UID: \"9fa8caf9-fab3-44d2-9841-593558e9a690\") " pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-gw8z5" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.496575 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-zwknl" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.506808 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-h9n4d" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.511430 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mjgl\" (UniqueName: \"kubernetes.io/projected/50963101-0a9e-453d-925b-ac6f08c545c4-kube-api-access-4mjgl\") pod \"heat-operator-controller-manager-69f49c598c-csfk2\" (UID: \"50963101-0a9e-453d-925b-ac6f08c545c4\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-csfk2" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.511458 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-z4jrr"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.512662 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-z4jrr" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.521180 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-8t7p2"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.522007 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-8t7p2" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.526232 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6bd4687957-vnb57"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.529836 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-qvdrn" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.530037 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-hthsr" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.534130 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-8t7p2"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.539346 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-qm5sw" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.540416 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-659dc6bbfc-qtr8c"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.541300 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-qtr8c" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.541847 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert\") pod \"infra-operator-controller-manager-79d975b745-s5ztn\" (UID: \"9a554278-1139-4942-904a-ffc60adde6de\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.541896 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztghv\" (UniqueName: \"kubernetes.io/projected/9a554278-1139-4942-904a-ffc60adde6de-kube-api-access-ztghv\") pod \"infra-operator-controller-manager-79d975b745-s5ztn\" (UID: \"9a554278-1139-4942-904a-ffc60adde6de\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" Feb 25 13:35:03 crc kubenswrapper[4815]: E0225 13:35:03.541952 4815 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 25 13:35:03 crc kubenswrapper[4815]: E0225 13:35:03.542022 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert podName:9a554278-1139-4942-904a-ffc60adde6de nodeName:}" failed. No retries permitted until 2026-02-25 13:35:04.042003787 +0000 UTC m=+861.843101841 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert") pod "infra-operator-controller-manager-79d975b745-s5ztn" (UID: "9a554278-1139-4942-904a-ffc60adde6de") : secret "infra-operator-webhook-server-cert" not found Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.541954 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqqrr\" (UniqueName: \"kubernetes.io/projected/c64a07be-3cda-473e-89b6-fd0347519c93-kube-api-access-xqqrr\") pod \"keystone-operator-controller-manager-b4d948c87-rx4fg\" (UID: \"c64a07be-3cda-473e-89b6-fd0347519c93\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-rx4fg" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.542268 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s79s\" (UniqueName: \"kubernetes.io/projected/12b1d896-ab7e-4097-afa4-791d39e1924e-kube-api-access-8s79s\") pod \"ironic-operator-controller-manager-554564d7fc-52psd\" (UID: \"12b1d896-ab7e-4097-afa4-791d39e1924e\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-52psd" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.542436 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96nkl\" (UniqueName: \"kubernetes.io/projected/b30720e2-30f0-480e-a44f-2b409be766c6-kube-api-access-96nkl\") pod \"nova-operator-controller-manager-567668f5cf-z4jrr\" (UID: \"b30720e2-30f0-480e-a44f-2b409be766c6\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-z4jrr" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.542791 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95kxh\" (UniqueName: \"kubernetes.io/projected/4dc8d07d-3e17-40a8-ac19-187729a64c4d-kube-api-access-95kxh\") pod \"manila-operator-controller-manager-67d996989d-5fq96\" (UID: \"4dc8d07d-3e17-40a8-ac19-187729a64c4d\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-5fq96" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.542924 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5mz2\" (UniqueName: \"kubernetes.io/projected/6bab4cd6-5eff-409c-83e3-ad5c1a6b6444-kube-api-access-q5mz2\") pod \"neutron-operator-controller-manager-6bd4687957-vnb57\" (UID: \"6bab4cd6-5eff-409c-83e3-ad5c1a6b6444\") " pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-vnb57" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.544901 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-lmnpr" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.551647 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-gw8z5" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.559115 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztghv\" (UniqueName: \"kubernetes.io/projected/9a554278-1139-4942-904a-ffc60adde6de-kube-api-access-ztghv\") pod \"infra-operator-controller-manager-79d975b745-s5ztn\" (UID: \"9a554278-1139-4942-904a-ffc60adde6de\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.559551 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s79s\" (UniqueName: \"kubernetes.io/projected/12b1d896-ab7e-4097-afa4-791d39e1924e-kube-api-access-8s79s\") pod \"ironic-operator-controller-manager-554564d7fc-52psd\" (UID: \"12b1d896-ab7e-4097-afa4-791d39e1924e\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-52psd" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.560241 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-659dc6bbfc-qtr8c"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.569694 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-z4jrr"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.593552 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.594317 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.596057 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-l8kdw" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.597687 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5955d8c787-62tw8"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.598534 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-62tw8" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.599312 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.599637 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-8852p" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.610152 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5955d8c787-62tw8"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.619593 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-9qrdg"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.620657 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9qrdg" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.626841 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.632310 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-wjmwg" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.632464 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-8vr5p"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.633481 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-8vr5p" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.636884 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-rjd8z" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.638184 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-csfk2" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.644329 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqqrr\" (UniqueName: \"kubernetes.io/projected/c64a07be-3cda-473e-89b6-fd0347519c93-kube-api-access-xqqrr\") pod \"keystone-operator-controller-manager-b4d948c87-rx4fg\" (UID: \"c64a07be-3cda-473e-89b6-fd0347519c93\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-rx4fg" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.644363 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7n4l\" (UniqueName: \"kubernetes.io/projected/43faa7ed-99a7-478c-8994-7a97333df231-kube-api-access-p7n4l\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p\" (UID: \"43faa7ed-99a7-478c-8994-7a97333df231\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.644396 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96nkl\" (UniqueName: \"kubernetes.io/projected/b30720e2-30f0-480e-a44f-2b409be766c6-kube-api-access-96nkl\") pod \"nova-operator-controller-manager-567668f5cf-z4jrr\" (UID: \"b30720e2-30f0-480e-a44f-2b409be766c6\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-z4jrr" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.644418 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx949\" (UniqueName: \"kubernetes.io/projected/b7cddc29-1c58-4edd-992f-165386b751ba-kube-api-access-cx949\") pod \"octavia-operator-controller-manager-659dc6bbfc-qtr8c\" (UID: \"b7cddc29-1c58-4edd-992f-165386b751ba\") " pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-qtr8c" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.644437 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95kxh\" (UniqueName: \"kubernetes.io/projected/4dc8d07d-3e17-40a8-ac19-187729a64c4d-kube-api-access-95kxh\") pod \"manila-operator-controller-manager-67d996989d-5fq96\" (UID: \"4dc8d07d-3e17-40a8-ac19-187729a64c4d\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-5fq96" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.644458 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5mz2\" (UniqueName: \"kubernetes.io/projected/6bab4cd6-5eff-409c-83e3-ad5c1a6b6444-kube-api-access-q5mz2\") pod \"neutron-operator-controller-manager-6bd4687957-vnb57\" (UID: \"6bab4cd6-5eff-409c-83e3-ad5c1a6b6444\") " pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-vnb57" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.644550 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q82ql\" (UniqueName: \"kubernetes.io/projected/697c602f-d6bc-478c-b89b-e3cc8aee4dcc-kube-api-access-q82ql\") pod \"ovn-operator-controller-manager-5955d8c787-62tw8\" (UID: \"697c602f-d6bc-478c-b89b-e3cc8aee4dcc\") " pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-62tw8" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.644569 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz6hj\" (UniqueName: \"kubernetes.io/projected/16cb48d6-2014-44ec-aaf5-6aef86d24cdf-kube-api-access-bz6hj\") pod \"mariadb-operator-controller-manager-6994f66f48-8t7p2\" (UID: \"16cb48d6-2014-44ec-aaf5-6aef86d24cdf\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-8t7p2" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.644605 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p\" (UID: \"43faa7ed-99a7-478c-8994-7a97333df231\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.644620 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj6cr\" (UniqueName: \"kubernetes.io/projected/3b1a8f42-91a7-46b3-96c6-0ef20a126678-kube-api-access-lj6cr\") pod \"placement-operator-controller-manager-8497b45c89-9qrdg\" (UID: \"3b1a8f42-91a7-46b3-96c6-0ef20a126678\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9qrdg" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.654977 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-9qrdg"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.655984 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-ccl5w" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.676444 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5mz2\" (UniqueName: \"kubernetes.io/projected/6bab4cd6-5eff-409c-83e3-ad5c1a6b6444-kube-api-access-q5mz2\") pod \"neutron-operator-controller-manager-6bd4687957-vnb57\" (UID: \"6bab4cd6-5eff-409c-83e3-ad5c1a6b6444\") " pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-vnb57" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.676599 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95kxh\" (UniqueName: \"kubernetes.io/projected/4dc8d07d-3e17-40a8-ac19-187729a64c4d-kube-api-access-95kxh\") pod \"manila-operator-controller-manager-67d996989d-5fq96\" (UID: \"4dc8d07d-3e17-40a8-ac19-187729a64c4d\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-5fq96" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.679478 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-8vr5p"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.703746 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96nkl\" (UniqueName: \"kubernetes.io/projected/b30720e2-30f0-480e-a44f-2b409be766c6-kube-api-access-96nkl\") pod \"nova-operator-controller-manager-567668f5cf-z4jrr\" (UID: \"b30720e2-30f0-480e-a44f-2b409be766c6\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-z4jrr" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.706472 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqqrr\" (UniqueName: \"kubernetes.io/projected/c64a07be-3cda-473e-89b6-fd0347519c93-kube-api-access-xqqrr\") pod \"keystone-operator-controller-manager-b4d948c87-rx4fg\" (UID: \"c64a07be-3cda-473e-89b6-fd0347519c93\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-rx4fg" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.714625 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-z4jrr" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.749182 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q82ql\" (UniqueName: \"kubernetes.io/projected/697c602f-d6bc-478c-b89b-e3cc8aee4dcc-kube-api-access-q82ql\") pod \"ovn-operator-controller-manager-5955d8c787-62tw8\" (UID: \"697c602f-d6bc-478c-b89b-e3cc8aee4dcc\") " pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-62tw8" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.749271 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz6hj\" (UniqueName: \"kubernetes.io/projected/16cb48d6-2014-44ec-aaf5-6aef86d24cdf-kube-api-access-bz6hj\") pod \"mariadb-operator-controller-manager-6994f66f48-8t7p2\" (UID: \"16cb48d6-2014-44ec-aaf5-6aef86d24cdf\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-8t7p2" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.749345 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9szd\" (UniqueName: \"kubernetes.io/projected/0673a594-2080-4325-80ce-f5597ad337c7-kube-api-access-q9szd\") pod \"swift-operator-controller-manager-68f46476f-8vr5p\" (UID: \"0673a594-2080-4325-80ce-f5597ad337c7\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-8vr5p" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.749394 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p\" (UID: \"43faa7ed-99a7-478c-8994-7a97333df231\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.749490 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj6cr\" (UniqueName: \"kubernetes.io/projected/3b1a8f42-91a7-46b3-96c6-0ef20a126678-kube-api-access-lj6cr\") pod \"placement-operator-controller-manager-8497b45c89-9qrdg\" (UID: \"3b1a8f42-91a7-46b3-96c6-0ef20a126678\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9qrdg" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.749598 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7n4l\" (UniqueName: \"kubernetes.io/projected/43faa7ed-99a7-478c-8994-7a97333df231-kube-api-access-p7n4l\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p\" (UID: \"43faa7ed-99a7-478c-8994-7a97333df231\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.749646 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx949\" (UniqueName: \"kubernetes.io/projected/b7cddc29-1c58-4edd-992f-165386b751ba-kube-api-access-cx949\") pod \"octavia-operator-controller-manager-659dc6bbfc-qtr8c\" (UID: \"b7cddc29-1c58-4edd-992f-165386b751ba\") " pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-qtr8c" Feb 25 13:35:03 crc kubenswrapper[4815]: E0225 13:35:03.750646 4815 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 13:35:03 crc kubenswrapper[4815]: E0225 13:35:03.750703 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert podName:43faa7ed-99a7-478c-8994-7a97333df231 nodeName:}" failed. No retries permitted until 2026-02-25 13:35:04.250686576 +0000 UTC m=+862.051784630 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" (UID: "43faa7ed-99a7-478c-8994-7a97333df231") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.781066 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-52psd" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.798772 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz6hj\" (UniqueName: \"kubernetes.io/projected/16cb48d6-2014-44ec-aaf5-6aef86d24cdf-kube-api-access-bz6hj\") pod \"mariadb-operator-controller-manager-6994f66f48-8t7p2\" (UID: \"16cb48d6-2014-44ec-aaf5-6aef86d24cdf\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-8t7p2" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.805464 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q82ql\" (UniqueName: \"kubernetes.io/projected/697c602f-d6bc-478c-b89b-e3cc8aee4dcc-kube-api-access-q82ql\") pod \"ovn-operator-controller-manager-5955d8c787-62tw8\" (UID: \"697c602f-d6bc-478c-b89b-e3cc8aee4dcc\") " pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-62tw8" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.806050 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx949\" (UniqueName: \"kubernetes.io/projected/b7cddc29-1c58-4edd-992f-165386b751ba-kube-api-access-cx949\") pod \"octavia-operator-controller-manager-659dc6bbfc-qtr8c\" (UID: \"b7cddc29-1c58-4edd-992f-165386b751ba\") " pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-qtr8c" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.832688 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-62tw8" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.833915 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj6cr\" (UniqueName: \"kubernetes.io/projected/3b1a8f42-91a7-46b3-96c6-0ef20a126678-kube-api-access-lj6cr\") pod \"placement-operator-controller-manager-8497b45c89-9qrdg\" (UID: \"3b1a8f42-91a7-46b3-96c6-0ef20a126678\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9qrdg" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.835030 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7n4l\" (UniqueName: \"kubernetes.io/projected/43faa7ed-99a7-478c-8994-7a97333df231-kube-api-access-p7n4l\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p\" (UID: \"43faa7ed-99a7-478c-8994-7a97333df231\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.847567 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-rx4fg" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.856652 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9szd\" (UniqueName: \"kubernetes.io/projected/0673a594-2080-4325-80ce-f5597ad337c7-kube-api-access-q9szd\") pod \"swift-operator-controller-manager-68f46476f-8vr5p\" (UID: \"0673a594-2080-4325-80ce-f5597ad337c7\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-8vr5p" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.863125 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-589c568786-rwzjh"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.864114 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-67d996989d-5fq96" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.866007 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-rwzjh" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.876160 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-5bf59" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.880756 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-vnb57" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.888302 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9szd\" (UniqueName: \"kubernetes.io/projected/0673a594-2080-4325-80ce-f5597ad337c7-kube-api-access-q9szd\") pod \"swift-operator-controller-manager-68f46476f-8vr5p\" (UID: \"0673a594-2080-4325-80ce-f5597ad337c7\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-8vr5p" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.899161 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-589c568786-rwzjh"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.950713 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5dc6794d5b-grtsf"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.951566 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-grtsf" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.953840 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-6pn4b" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.957569 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5dc6794d5b-grtsf"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.958336 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8vc6\" (UniqueName: \"kubernetes.io/projected/136c32bc-0a85-4a36-b527-a787fa7158f1-kube-api-access-w8vc6\") pod \"telemetry-operator-controller-manager-589c568786-rwzjh\" (UID: \"136c32bc-0a85-4a36-b527-a787fa7158f1\") " pod="openstack-operators/telemetry-operator-controller-manager-589c568786-rwzjh" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.965155 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-bccc79885-v825g"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.966003 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-v825g" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.970774 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-mrn47" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.971038 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9qrdg" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.975137 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-bccc79885-v825g"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.975483 4815 generic.go:334] "Generic (PLEG): container finished" podID="59dc6b8b-e3e2-4f25-91bf-310d17bbd603" containerID="e687322c626dafb096300a9c8ba7e52208ef18d15312da9bcd8afd9d613c2ff3" exitCode=0 Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.975521 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tddkm" event={"ID":"59dc6b8b-e3e2-4f25-91bf-310d17bbd603","Type":"ContainerDied","Data":"e687322c626dafb096300a9c8ba7e52208ef18d15312da9bcd8afd9d613c2ff3"} Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.990496 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h"] Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.991435 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.996679 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.997269 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-mffwd" Feb 25 13:35:03 crc kubenswrapper[4815]: I0225 13:35:03.997471 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.005611 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h"] Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.032569 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7l95p"] Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.033318 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7l95p"] Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.033396 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7l95p" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.042664 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-wkp7m" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.043053 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-8vr5p" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.043952 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-8t7p2" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.047568 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-qtr8c" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.059855 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.059934 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhh7w\" (UniqueName: \"kubernetes.io/projected/486a2442-b575-491d-9c18-4185e5a647fe-kube-api-access-hhh7w\") pod \"watcher-operator-controller-manager-bccc79885-v825g\" (UID: \"486a2442-b575-491d-9c18-4185e5a647fe\") " pod="openstack-operators/watcher-operator-controller-manager-bccc79885-v825g" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.059963 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.060027 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2mpv\" (UniqueName: \"kubernetes.io/projected/22b232e5-5045-404a-a248-01babe4886c3-kube-api-access-c2mpv\") pod \"rabbitmq-cluster-operator-manager-668c99d594-7l95p\" (UID: \"22b232e5-5045-404a-a248-01babe4886c3\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7l95p" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.060170 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t8lz\" (UniqueName: \"kubernetes.io/projected/509894d6-7814-4f85-9eed-f1946b029dbf-kube-api-access-2t8lz\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.060270 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert\") pod \"infra-operator-controller-manager-79d975b745-s5ztn\" (UID: \"9a554278-1139-4942-904a-ffc60adde6de\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.060296 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf8d2\" (UniqueName: \"kubernetes.io/projected/21de3543-3d6a-4adf-b104-f9fcaee4532b-kube-api-access-tf8d2\") pod \"test-operator-controller-manager-5dc6794d5b-grtsf\" (UID: \"21de3543-3d6a-4adf-b104-f9fcaee4532b\") " pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-grtsf" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.060358 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8vc6\" (UniqueName: \"kubernetes.io/projected/136c32bc-0a85-4a36-b527-a787fa7158f1-kube-api-access-w8vc6\") pod \"telemetry-operator-controller-manager-589c568786-rwzjh\" (UID: \"136c32bc-0a85-4a36-b527-a787fa7158f1\") " pod="openstack-operators/telemetry-operator-controller-manager-589c568786-rwzjh" Feb 25 13:35:04 crc kubenswrapper[4815]: E0225 13:35:04.061339 4815 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 25 13:35:04 crc kubenswrapper[4815]: E0225 13:35:04.061605 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert podName:9a554278-1139-4942-904a-ffc60adde6de nodeName:}" failed. No retries permitted until 2026-02-25 13:35:05.061590165 +0000 UTC m=+862.862688219 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert") pod "infra-operator-controller-manager-79d975b745-s5ztn" (UID: "9a554278-1139-4942-904a-ffc60adde6de") : secret "infra-operator-webhook-server-cert" not found Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.080274 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8vc6\" (UniqueName: \"kubernetes.io/projected/136c32bc-0a85-4a36-b527-a787fa7158f1-kube-api-access-w8vc6\") pod \"telemetry-operator-controller-manager-589c568786-rwzjh\" (UID: \"136c32bc-0a85-4a36-b527-a787fa7158f1\") " pod="openstack-operators/telemetry-operator-controller-manager-589c568786-rwzjh" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.090120 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-rwzjh" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.125107 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-zwknl"] Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.153791 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t9pdn"] Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.154005 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t9pdn" podUID="7c43e580-fd0a-455b-92e9-c87d13dccccc" containerName="registry-server" containerID="cri-o://477af7a0fea6b9b41dee24877149c9d883dfb2982c91507b6d577f142e3ad967" gracePeriod=2 Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.162733 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2mpv\" (UniqueName: \"kubernetes.io/projected/22b232e5-5045-404a-a248-01babe4886c3-kube-api-access-c2mpv\") pod \"rabbitmq-cluster-operator-manager-668c99d594-7l95p\" (UID: \"22b232e5-5045-404a-a248-01babe4886c3\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7l95p" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.162783 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t8lz\" (UniqueName: \"kubernetes.io/projected/509894d6-7814-4f85-9eed-f1946b029dbf-kube-api-access-2t8lz\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.162847 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf8d2\" (UniqueName: \"kubernetes.io/projected/21de3543-3d6a-4adf-b104-f9fcaee4532b-kube-api-access-tf8d2\") pod \"test-operator-controller-manager-5dc6794d5b-grtsf\" (UID: \"21de3543-3d6a-4adf-b104-f9fcaee4532b\") " pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-grtsf" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.163595 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.163631 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhh7w\" (UniqueName: \"kubernetes.io/projected/486a2442-b575-491d-9c18-4185e5a647fe-kube-api-access-hhh7w\") pod \"watcher-operator-controller-manager-bccc79885-v825g\" (UID: \"486a2442-b575-491d-9c18-4185e5a647fe\") " pod="openstack-operators/watcher-operator-controller-manager-bccc79885-v825g" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.163651 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:04 crc kubenswrapper[4815]: E0225 13:35:04.163749 4815 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 25 13:35:04 crc kubenswrapper[4815]: E0225 13:35:04.163750 4815 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 25 13:35:04 crc kubenswrapper[4815]: E0225 13:35:04.163791 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs podName:509894d6-7814-4f85-9eed-f1946b029dbf nodeName:}" failed. No retries permitted until 2026-02-25 13:35:04.663776764 +0000 UTC m=+862.464874808 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs") pod "openstack-operator-controller-manager-5f47dfdb6c-slw4h" (UID: "509894d6-7814-4f85-9eed-f1946b029dbf") : secret "metrics-server-cert" not found Feb 25 13:35:04 crc kubenswrapper[4815]: E0225 13:35:04.163803 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs podName:509894d6-7814-4f85-9eed-f1946b029dbf nodeName:}" failed. No retries permitted until 2026-02-25 13:35:04.663797905 +0000 UTC m=+862.464895959 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs") pod "openstack-operator-controller-manager-5f47dfdb6c-slw4h" (UID: "509894d6-7814-4f85-9eed-f1946b029dbf") : secret "webhook-server-cert" not found Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.190999 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhh7w\" (UniqueName: \"kubernetes.io/projected/486a2442-b575-491d-9c18-4185e5a647fe-kube-api-access-hhh7w\") pod \"watcher-operator-controller-manager-bccc79885-v825g\" (UID: \"486a2442-b575-491d-9c18-4185e5a647fe\") " pod="openstack-operators/watcher-operator-controller-manager-bccc79885-v825g" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.191520 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf8d2\" (UniqueName: \"kubernetes.io/projected/21de3543-3d6a-4adf-b104-f9fcaee4532b-kube-api-access-tf8d2\") pod \"test-operator-controller-manager-5dc6794d5b-grtsf\" (UID: \"21de3543-3d6a-4adf-b104-f9fcaee4532b\") " pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-grtsf" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.192140 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t8lz\" (UniqueName: \"kubernetes.io/projected/509894d6-7814-4f85-9eed-f1946b029dbf-kube-api-access-2t8lz\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.194533 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2mpv\" (UniqueName: \"kubernetes.io/projected/22b232e5-5045-404a-a248-01babe4886c3-kube-api-access-c2mpv\") pod \"rabbitmq-cluster-operator-manager-668c99d594-7l95p\" (UID: \"22b232e5-5045-404a-a248-01babe4886c3\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7l95p" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.265103 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p\" (UID: \"43faa7ed-99a7-478c-8994-7a97333df231\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" Feb 25 13:35:04 crc kubenswrapper[4815]: E0225 13:35:04.265664 4815 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 13:35:04 crc kubenswrapper[4815]: E0225 13:35:04.265702 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert podName:43faa7ed-99a7-478c-8994-7a97333df231 nodeName:}" failed. No retries permitted until 2026-02-25 13:35:05.265688673 +0000 UTC m=+863.066786727 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" (UID: "43faa7ed-99a7-478c-8994-7a97333df231") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.414872 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-grtsf" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.434177 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-h9n4d"] Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.441999 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-v825g" Feb 25 13:35:04 crc kubenswrapper[4815]: W0225 13:35:04.460088 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbace17a8_ce7d_4df9_8259_68931eeecde8.slice/crio-6c3326abd0badd4d5e55822bd65038240e00259f9c208fda6e1cdbf1343c5047 WatchSource:0}: Error finding container 6c3326abd0badd4d5e55822bd65038240e00259f9c208fda6e1cdbf1343c5047: Status 404 returned error can't find the container with id 6c3326abd0badd4d5e55822bd65038240e00259f9c208fda6e1cdbf1343c5047 Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.470348 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-qm5sw"] Feb 25 13:35:04 crc kubenswrapper[4815]: W0225 13:35:04.478610 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4956173_674d_4b10_9a3f_9996f9190299.slice/crio-a44a1db624c470174944094670d7621f00610a0083702fd27410d83ead5f04f4 WatchSource:0}: Error finding container a44a1db624c470174944094670d7621f00610a0083702fd27410d83ead5f04f4: Status 404 returned error can't find the container with id a44a1db624c470174944094670d7621f00610a0083702fd27410d83ead5f04f4 Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.494764 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7l95p" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.665011 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-784b5bb6c5-gw8z5"] Feb 25 13:35:04 crc kubenswrapper[4815]: W0225 13:35:04.668515 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fa8caf9_fab3_44d2_9841_593558e9a690.slice/crio-43d0b2d09f52ade93f80ce495123b16c49539454e0a1084120bc58d82ba5a877 WatchSource:0}: Error finding container 43d0b2d09f52ade93f80ce495123b16c49539454e0a1084120bc58d82ba5a877: Status 404 returned error can't find the container with id 43d0b2d09f52ade93f80ce495123b16c49539454e0a1084120bc58d82ba5a877 Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.672042 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.672085 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:04 crc kubenswrapper[4815]: E0225 13:35:04.672240 4815 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 25 13:35:04 crc kubenswrapper[4815]: E0225 13:35:04.672293 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs podName:509894d6-7814-4f85-9eed-f1946b029dbf nodeName:}" failed. No retries permitted until 2026-02-25 13:35:05.67227638 +0000 UTC m=+863.473374434 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs") pod "openstack-operator-controller-manager-5f47dfdb6c-slw4h" (UID: "509894d6-7814-4f85-9eed-f1946b029dbf") : secret "metrics-server-cert" not found Feb 25 13:35:04 crc kubenswrapper[4815]: E0225 13:35:04.672642 4815 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 25 13:35:04 crc kubenswrapper[4815]: E0225 13:35:04.672719 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs podName:509894d6-7814-4f85-9eed-f1946b029dbf nodeName:}" failed. No retries permitted until 2026-02-25 13:35:05.672698653 +0000 UTC m=+863.473796717 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs") pod "openstack-operator-controller-manager-5f47dfdb6c-slw4h" (UID: "509894d6-7814-4f85-9eed-f1946b029dbf") : secret "webhook-server-cert" not found Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.708272 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-csfk2"] Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.746583 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-ccl5w"] Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.762126 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-z4jrr"] Feb 25 13:35:04 crc kubenswrapper[4815]: W0225 13:35:04.764152 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb30720e2_30f0_480e_a44f_2b409be766c6.slice/crio-9c972d6b02de9de2bc6ae0b0bd8961391826b08876c24dd9a850a16b65cb7421 WatchSource:0}: Error finding container 9c972d6b02de9de2bc6ae0b0bd8961391826b08876c24dd9a850a16b65cb7421: Status 404 returned error can't find the container with id 9c972d6b02de9de2bc6ae0b0bd8961391826b08876c24dd9a850a16b65cb7421 Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.956194 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t9pdn" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.976836 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c43e580-fd0a-455b-92e9-c87d13dccccc-catalog-content\") pod \"7c43e580-fd0a-455b-92e9-c87d13dccccc\" (UID: \"7c43e580-fd0a-455b-92e9-c87d13dccccc\") " Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.983147 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgsp2\" (UniqueName: \"kubernetes.io/projected/7c43e580-fd0a-455b-92e9-c87d13dccccc-kube-api-access-mgsp2\") pod \"7c43e580-fd0a-455b-92e9-c87d13dccccc\" (UID: \"7c43e580-fd0a-455b-92e9-c87d13dccccc\") " Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.983180 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c43e580-fd0a-455b-92e9-c87d13dccccc-utilities\") pod \"7c43e580-fd0a-455b-92e9-c87d13dccccc\" (UID: \"7c43e580-fd0a-455b-92e9-c87d13dccccc\") " Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.984149 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c43e580-fd0a-455b-92e9-c87d13dccccc-utilities" (OuterVolumeSpecName: "utilities") pod "7c43e580-fd0a-455b-92e9-c87d13dccccc" (UID: "7c43e580-fd0a-455b-92e9-c87d13dccccc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.984780 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-h9n4d" event={"ID":"bace17a8-ce7d-4df9-8259-68931eeecde8","Type":"ContainerStarted","Data":"6c3326abd0badd4d5e55822bd65038240e00259f9c208fda6e1cdbf1343c5047"} Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.985851 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-csfk2" event={"ID":"50963101-0a9e-453d-925b-ac6f08c545c4","Type":"ContainerStarted","Data":"d116215131924e63cbf7e7fe3306959bee8037cab7120417ec977119efc613d3"} Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.988014 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c43e580-fd0a-455b-92e9-c87d13dccccc-kube-api-access-mgsp2" (OuterVolumeSpecName: "kube-api-access-mgsp2") pod "7c43e580-fd0a-455b-92e9-c87d13dccccc" (UID: "7c43e580-fd0a-455b-92e9-c87d13dccccc"). InnerVolumeSpecName "kube-api-access-mgsp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.991117 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-zwknl" event={"ID":"62c00caf-05cf-4229-9349-341a2b597b12","Type":"ContainerStarted","Data":"3197d223e351a89bfb7cd7d8d823ae5ac1490c195e2c82825db00836021d448f"} Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.995055 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tddkm" event={"ID":"59dc6b8b-e3e2-4f25-91bf-310d17bbd603","Type":"ContainerStarted","Data":"d419ac199f099c4ac254ef71a5cdaa1d318d0547c94f779cd8aa72515ebb27ad"} Feb 25 13:35:04 crc kubenswrapper[4815]: I0225 13:35:04.997796 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-qm5sw" event={"ID":"c4956173-674d-4b10-9a3f-9996f9190299","Type":"ContainerStarted","Data":"a44a1db624c470174944094670d7621f00610a0083702fd27410d83ead5f04f4"} Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.001084 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c43e580-fd0a-455b-92e9-c87d13dccccc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c43e580-fd0a-455b-92e9-c87d13dccccc" (UID: "7c43e580-fd0a-455b-92e9-c87d13dccccc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.003698 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-ccl5w" event={"ID":"33e59af6-b9bb-4b55-b6a9-c6963e269b94","Type":"ContainerStarted","Data":"93df9ab86a7ededaf2a49a59cd90fe9fa8275e88dade7cc5f3bfaa65cd7c325f"} Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.004537 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-gw8z5" event={"ID":"9fa8caf9-fab3-44d2-9841-593558e9a690","Type":"ContainerStarted","Data":"43d0b2d09f52ade93f80ce495123b16c49539454e0a1084120bc58d82ba5a877"} Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.005623 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-z4jrr" event={"ID":"b30720e2-30f0-480e-a44f-2b409be766c6","Type":"ContainerStarted","Data":"9c972d6b02de9de2bc6ae0b0bd8961391826b08876c24dd9a850a16b65cb7421"} Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.010122 4815 generic.go:334] "Generic (PLEG): container finished" podID="7c43e580-fd0a-455b-92e9-c87d13dccccc" containerID="477af7a0fea6b9b41dee24877149c9d883dfb2982c91507b6d577f142e3ad967" exitCode=0 Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.010155 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t9pdn" event={"ID":"7c43e580-fd0a-455b-92e9-c87d13dccccc","Type":"ContainerDied","Data":"477af7a0fea6b9b41dee24877149c9d883dfb2982c91507b6d577f142e3ad967"} Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.010173 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t9pdn" event={"ID":"7c43e580-fd0a-455b-92e9-c87d13dccccc","Type":"ContainerDied","Data":"3c016f8ce591eb0dbe32a117b723b8e489c48b62b5754e122e2ddbb83d004bf6"} Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.010176 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t9pdn" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.010188 4815 scope.go:117] "RemoveContainer" containerID="477af7a0fea6b9b41dee24877149c9d883dfb2982c91507b6d577f142e3ad967" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.014712 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tddkm" podStartSLOduration=2.379951582 podStartE2EDuration="5.014701823s" podCreationTimestamp="2026-02-25 13:35:00 +0000 UTC" firstStartedPulling="2026-02-25 13:35:01.913038871 +0000 UTC m=+859.714136925" lastFinishedPulling="2026-02-25 13:35:04.547789112 +0000 UTC m=+862.348887166" observedRunningTime="2026-02-25 13:35:05.01396878 +0000 UTC m=+862.815066844" watchObservedRunningTime="2026-02-25 13:35:05.014701823 +0000 UTC m=+862.815799877" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.034699 4815 scope.go:117] "RemoveContainer" containerID="5ec8c649f5cd3841cf2919eaf8fe6f595bcb04e2a6bee85c5e451d2dce16d6af" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.042236 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t9pdn"] Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.046718 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t9pdn"] Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.060907 4815 scope.go:117] "RemoveContainer" containerID="46850230513cda51c82c76d9278dd5a7c6b196cf196f46d01bc8adc6b10eb94c" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.079842 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-52psd"] Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.084702 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert\") pod \"infra-operator-controller-manager-79d975b745-s5ztn\" (UID: \"9a554278-1139-4942-904a-ffc60adde6de\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.084866 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c43e580-fd0a-455b-92e9-c87d13dccccc-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.084877 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgsp2\" (UniqueName: \"kubernetes.io/projected/7c43e580-fd0a-455b-92e9-c87d13dccccc-kube-api-access-mgsp2\") on node \"crc\" DevicePath \"\"" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.084887 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c43e580-fd0a-455b-92e9-c87d13dccccc-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.085143 4815 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.085214 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert podName:9a554278-1139-4942-904a-ffc60adde6de nodeName:}" failed. No retries permitted until 2026-02-25 13:35:07.085193081 +0000 UTC m=+864.886291195 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert") pod "infra-operator-controller-manager-79d975b745-s5ztn" (UID: "9a554278-1139-4942-904a-ffc60adde6de") : secret "infra-operator-webhook-server-cert" not found Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.097313 4815 scope.go:117] "RemoveContainer" containerID="477af7a0fea6b9b41dee24877149c9d883dfb2982c91507b6d577f142e3ad967" Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.099189 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"477af7a0fea6b9b41dee24877149c9d883dfb2982c91507b6d577f142e3ad967\": container with ID starting with 477af7a0fea6b9b41dee24877149c9d883dfb2982c91507b6d577f142e3ad967 not found: ID does not exist" containerID="477af7a0fea6b9b41dee24877149c9d883dfb2982c91507b6d577f142e3ad967" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.099219 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"477af7a0fea6b9b41dee24877149c9d883dfb2982c91507b6d577f142e3ad967"} err="failed to get container status \"477af7a0fea6b9b41dee24877149c9d883dfb2982c91507b6d577f142e3ad967\": rpc error: code = NotFound desc = could not find container \"477af7a0fea6b9b41dee24877149c9d883dfb2982c91507b6d577f142e3ad967\": container with ID starting with 477af7a0fea6b9b41dee24877149c9d883dfb2982c91507b6d577f142e3ad967 not found: ID does not exist" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.099245 4815 scope.go:117] "RemoveContainer" containerID="5ec8c649f5cd3841cf2919eaf8fe6f595bcb04e2a6bee85c5e451d2dce16d6af" Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.099604 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ec8c649f5cd3841cf2919eaf8fe6f595bcb04e2a6bee85c5e451d2dce16d6af\": container with ID starting with 5ec8c649f5cd3841cf2919eaf8fe6f595bcb04e2a6bee85c5e451d2dce16d6af not found: ID does not exist" containerID="5ec8c649f5cd3841cf2919eaf8fe6f595bcb04e2a6bee85c5e451d2dce16d6af" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.099631 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ec8c649f5cd3841cf2919eaf8fe6f595bcb04e2a6bee85c5e451d2dce16d6af"} err="failed to get container status \"5ec8c649f5cd3841cf2919eaf8fe6f595bcb04e2a6bee85c5e451d2dce16d6af\": rpc error: code = NotFound desc = could not find container \"5ec8c649f5cd3841cf2919eaf8fe6f595bcb04e2a6bee85c5e451d2dce16d6af\": container with ID starting with 5ec8c649f5cd3841cf2919eaf8fe6f595bcb04e2a6bee85c5e451d2dce16d6af not found: ID does not exist" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.099649 4815 scope.go:117] "RemoveContainer" containerID="46850230513cda51c82c76d9278dd5a7c6b196cf196f46d01bc8adc6b10eb94c" Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.103134 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46850230513cda51c82c76d9278dd5a7c6b196cf196f46d01bc8adc6b10eb94c\": container with ID starting with 46850230513cda51c82c76d9278dd5a7c6b196cf196f46d01bc8adc6b10eb94c not found: ID does not exist" containerID="46850230513cda51c82c76d9278dd5a7c6b196cf196f46d01bc8adc6b10eb94c" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.103221 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46850230513cda51c82c76d9278dd5a7c6b196cf196f46d01bc8adc6b10eb94c"} err="failed to get container status \"46850230513cda51c82c76d9278dd5a7c6b196cf196f46d01bc8adc6b10eb94c\": rpc error: code = NotFound desc = could not find container \"46850230513cda51c82c76d9278dd5a7c6b196cf196f46d01bc8adc6b10eb94c\": container with ID starting with 46850230513cda51c82c76d9278dd5a7c6b196cf196f46d01bc8adc6b10eb94c not found: ID does not exist" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.112491 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6bd4687957-vnb57"] Feb 25 13:35:05 crc kubenswrapper[4815]: W0225 13:35:05.112685 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12b1d896_ab7e_4097_afa4_791d39e1924e.slice/crio-ea194e56f07d92a715e2143a6a23679a0f887f3da0a461293b4ae11f148786cf WatchSource:0}: Error finding container ea194e56f07d92a715e2143a6a23679a0f887f3da0a461293b4ae11f148786cf: Status 404 returned error can't find the container with id ea194e56f07d92a715e2143a6a23679a0f887f3da0a461293b4ae11f148786cf Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.132138 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-659dc6bbfc-qtr8c"] Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.145410 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-9qrdg"] Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.150997 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-5fq96"] Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.155093 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5955d8c787-62tw8"] Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.165684 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-589c568786-rwzjh"] Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.174140 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-rx4fg"] Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.179490 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-8t7p2"] Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.187348 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-8vr5p"] Feb 25 13:35:05 crc kubenswrapper[4815]: W0225 13:35:05.189909 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod697c602f_d6bc_478c_b89b_e3cc8aee4dcc.slice/crio-89e2d790602597843fb7ae8cdf54796ebe449f234f857debbca5fdca04e9fa7f WatchSource:0}: Error finding container 89e2d790602597843fb7ae8cdf54796ebe449f234f857debbca5fdca04e9fa7f: Status 404 returned error can't find the container with id 89e2d790602597843fb7ae8cdf54796ebe449f234f857debbca5fdca04e9fa7f Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.196320 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-bccc79885-v825g"] Feb 25 13:35:05 crc kubenswrapper[4815]: W0225 13:35:05.203963 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0673a594_2080_4325_80ce_f5597ad337c7.slice/crio-62373defac8d7c8e364c47e8cc21583fb1ecfd84860702c119634e00e404e131 WatchSource:0}: Error finding container 62373defac8d7c8e364c47e8cc21583fb1ecfd84860702c119634e00e404e131: Status 404 returned error can't find the container with id 62373defac8d7c8e364c47e8cc21583fb1ecfd84860702c119634e00e404e131 Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.204020 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5dc6794d5b-grtsf"] Feb 25 13:35:05 crc kubenswrapper[4815]: W0225 13:35:05.206332 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16cb48d6_2014_44ec_aaf5_6aef86d24cdf.slice/crio-50272a6f57761cbe132b52ded8c8b616b0a374fd4d43bf61db1ff07e622423dc WatchSource:0}: Error finding container 50272a6f57761cbe132b52ded8c8b616b0a374fd4d43bf61db1ff07e622423dc: Status 404 returned error can't find the container with id 50272a6f57761cbe132b52ded8c8b616b0a374fd4d43bf61db1ff07e622423dc Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.209587 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:f4143497c70c048a7733c284060347a0c74ef4e628aca22ee191e5bc9e4c7192,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q82ql,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-5955d8c787-62tw8_openstack-operators(697c602f-d6bc-478c-b89b-e3cc8aee4dcc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.210712 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:f1158ec4d879c4646eee4323bc501eba4d377beb2ad6fbe08ed30070c441ac26,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-95kxh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-67d996989d-5fq96_openstack-operators(4dc8d07d-3e17-40a8-ac19-187729a64c4d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.210834 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7l95p"] Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.212235 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-67d996989d-5fq96" podUID="4dc8d07d-3e17-40a8-ac19-187729a64c4d" Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.212267 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-62tw8" podUID="697c602f-d6bc-478c-b89b-e3cc8aee4dcc" Feb 25 13:35:05 crc kubenswrapper[4815]: W0225 13:35:05.219782 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod486a2442_b575_491d_9c18_4185e5a647fe.slice/crio-e75cf2629b310a47e0944ed937e43830a1ca806b6191ba701394f09c5188d1f0 WatchSource:0}: Error finding container e75cf2629b310a47e0944ed937e43830a1ca806b6191ba701394f09c5188d1f0: Status 404 returned error can't find the container with id e75cf2629b310a47e0944ed937e43830a1ca806b6191ba701394f09c5188d1f0 Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.220045 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:a18f12497b7159b100fcfd72c7ba2273d0669a5c00600a9ff1333bca028f256a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bz6hj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-6994f66f48-8t7p2_openstack-operators(16cb48d6-2014-44ec-aaf5-6aef86d24cdf): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.221226 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-8t7p2" podUID="16cb48d6-2014-44ec-aaf5-6aef86d24cdf" Feb 25 13:35:05 crc kubenswrapper[4815]: W0225 13:35:05.223864 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21de3543_3d6a_4adf_b104_f9fcaee4532b.slice/crio-dca0c5aea4158095d5ea1f5789c1903c5f186248e7c9a044af25441ce23a730d WatchSource:0}: Error finding container dca0c5aea4158095d5ea1f5789c1903c5f186248e7c9a044af25441ce23a730d: Status 404 returned error can't find the container with id dca0c5aea4158095d5ea1f5789c1903c5f186248e7c9a044af25441ce23a730d Feb 25 13:35:05 crc kubenswrapper[4815]: W0225 13:35:05.228715 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22b232e5_5045_404a_a248_01babe4886c3.slice/crio-131c02e492153c47b6ccb2975508ee129d16581e9d87e77a3ec2d83c8d9b2d06 WatchSource:0}: Error finding container 131c02e492153c47b6ccb2975508ee129d16581e9d87e77a3ec2d83c8d9b2d06: Status 404 returned error can't find the container with id 131c02e492153c47b6ccb2975508ee129d16581e9d87e77a3ec2d83c8d9b2d06 Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.234067 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c2mpv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-7l95p_openstack-operators(22b232e5-5045-404a-a248-01babe4886c3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.234067 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:38e6a5bd24ab1684f22a64186fe99a7cdc7897eb7feb715ec1704eea7596dd98,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tf8d2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5dc6794d5b-grtsf_openstack-operators(21de3543-3d6a-4adf-b104-f9fcaee4532b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.234067 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:06311600a491c689493552e7ff26e36df740fa4e7c143fca874bef19f24afb97,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hhh7w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-bccc79885-v825g_openstack-operators(486a2442-b575-491d-9c18-4185e5a647fe): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.235245 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-v825g" podUID="486a2442-b575-491d-9c18-4185e5a647fe" Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.235271 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-grtsf" podUID="21de3543-3d6a-4adf-b104-f9fcaee4532b" Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.235260 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7l95p" podUID="22b232e5-5045-404a-a248-01babe4886c3" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.288185 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p\" (UID: \"43faa7ed-99a7-478c-8994-7a97333df231\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.288600 4815 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.288732 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert podName:43faa7ed-99a7-478c-8994-7a97333df231 nodeName:}" failed. No retries permitted until 2026-02-25 13:35:07.288709052 +0000 UTC m=+865.089807106 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" (UID: "43faa7ed-99a7-478c-8994-7a97333df231") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.699880 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:05 crc kubenswrapper[4815]: I0225 13:35:05.699949 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.700266 4815 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.700326 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs podName:509894d6-7814-4f85-9eed-f1946b029dbf nodeName:}" failed. No retries permitted until 2026-02-25 13:35:07.700305893 +0000 UTC m=+865.501403947 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs") pod "openstack-operator-controller-manager-5f47dfdb6c-slw4h" (UID: "509894d6-7814-4f85-9eed-f1946b029dbf") : secret "metrics-server-cert" not found Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.700378 4815 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 25 13:35:05 crc kubenswrapper[4815]: E0225 13:35:05.700401 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs podName:509894d6-7814-4f85-9eed-f1946b029dbf nodeName:}" failed. No retries permitted until 2026-02-25 13:35:07.700392526 +0000 UTC m=+865.501490580 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs") pod "openstack-operator-controller-manager-5f47dfdb6c-slw4h" (UID: "509894d6-7814-4f85-9eed-f1946b029dbf") : secret "webhook-server-cert" not found Feb 25 13:35:06 crc kubenswrapper[4815]: I0225 13:35:06.053336 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-62tw8" event={"ID":"697c602f-d6bc-478c-b89b-e3cc8aee4dcc","Type":"ContainerStarted","Data":"89e2d790602597843fb7ae8cdf54796ebe449f234f857debbca5fdca04e9fa7f"} Feb 25 13:35:06 crc kubenswrapper[4815]: E0225 13:35:06.058742 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f4143497c70c048a7733c284060347a0c74ef4e628aca22ee191e5bc9e4c7192\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-62tw8" podUID="697c602f-d6bc-478c-b89b-e3cc8aee4dcc" Feb 25 13:35:06 crc kubenswrapper[4815]: I0225 13:35:06.069581 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-grtsf" event={"ID":"21de3543-3d6a-4adf-b104-f9fcaee4532b","Type":"ContainerStarted","Data":"dca0c5aea4158095d5ea1f5789c1903c5f186248e7c9a044af25441ce23a730d"} Feb 25 13:35:06 crc kubenswrapper[4815]: E0225 13:35:06.076735 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:38e6a5bd24ab1684f22a64186fe99a7cdc7897eb7feb715ec1704eea7596dd98\\\"\"" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-grtsf" podUID="21de3543-3d6a-4adf-b104-f9fcaee4532b" Feb 25 13:35:06 crc kubenswrapper[4815]: I0225 13:35:06.093567 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-52psd" event={"ID":"12b1d896-ab7e-4097-afa4-791d39e1924e","Type":"ContainerStarted","Data":"ea194e56f07d92a715e2143a6a23679a0f887f3da0a461293b4ae11f148786cf"} Feb 25 13:35:06 crc kubenswrapper[4815]: I0225 13:35:06.106113 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7l95p" event={"ID":"22b232e5-5045-404a-a248-01babe4886c3","Type":"ContainerStarted","Data":"131c02e492153c47b6ccb2975508ee129d16581e9d87e77a3ec2d83c8d9b2d06"} Feb 25 13:35:06 crc kubenswrapper[4815]: E0225 13:35:06.116461 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7l95p" podUID="22b232e5-5045-404a-a248-01babe4886c3" Feb 25 13:35:06 crc kubenswrapper[4815]: I0225 13:35:06.132125 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-qtr8c" event={"ID":"b7cddc29-1c58-4edd-992f-165386b751ba","Type":"ContainerStarted","Data":"611fdc15d40c93d908d49a8214544718fe03197fd058f6b58e4ee0214c810c26"} Feb 25 13:35:06 crc kubenswrapper[4815]: I0225 13:35:06.134067 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-8vr5p" event={"ID":"0673a594-2080-4325-80ce-f5597ad337c7","Type":"ContainerStarted","Data":"62373defac8d7c8e364c47e8cc21583fb1ecfd84860702c119634e00e404e131"} Feb 25 13:35:06 crc kubenswrapper[4815]: I0225 13:35:06.137239 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-vnb57" event={"ID":"6bab4cd6-5eff-409c-83e3-ad5c1a6b6444","Type":"ContainerStarted","Data":"372a125e9d37274dc3220a098837fc5b01aa0abd56b83098ec52ee9610b67060"} Feb 25 13:35:06 crc kubenswrapper[4815]: I0225 13:35:06.139923 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-rwzjh" event={"ID":"136c32bc-0a85-4a36-b527-a787fa7158f1","Type":"ContainerStarted","Data":"8b0198a43bd31440152026e577b15b4413b97c1afcec6d5fd48a2c8e02b010db"} Feb 25 13:35:06 crc kubenswrapper[4815]: I0225 13:35:06.141106 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-rx4fg" event={"ID":"c64a07be-3cda-473e-89b6-fd0347519c93","Type":"ContainerStarted","Data":"a249952cdc9ef97d1be56ffd2ece5dee4d551d58b2b52e344cf230733b43c490"} Feb 25 13:35:06 crc kubenswrapper[4815]: I0225 13:35:06.145572 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-67d996989d-5fq96" event={"ID":"4dc8d07d-3e17-40a8-ac19-187729a64c4d","Type":"ContainerStarted","Data":"a4b310a37502e37adba9df51feb18543023645b64c554f48260c8ddf54f0e3f1"} Feb 25 13:35:06 crc kubenswrapper[4815]: E0225 13:35:06.153826 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:f1158ec4d879c4646eee4323bc501eba4d377beb2ad6fbe08ed30070c441ac26\\\"\"" pod="openstack-operators/manila-operator-controller-manager-67d996989d-5fq96" podUID="4dc8d07d-3e17-40a8-ac19-187729a64c4d" Feb 25 13:35:06 crc kubenswrapper[4815]: I0225 13:35:06.158741 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9qrdg" event={"ID":"3b1a8f42-91a7-46b3-96c6-0ef20a126678","Type":"ContainerStarted","Data":"0e17a83c39a1969f35247911e2884b02a4960296c93a8f4a7131686f51213db7"} Feb 25 13:35:06 crc kubenswrapper[4815]: I0225 13:35:06.171983 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-v825g" event={"ID":"486a2442-b575-491d-9c18-4185e5a647fe","Type":"ContainerStarted","Data":"e75cf2629b310a47e0944ed937e43830a1ca806b6191ba701394f09c5188d1f0"} Feb 25 13:35:06 crc kubenswrapper[4815]: I0225 13:35:06.174309 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-8t7p2" event={"ID":"16cb48d6-2014-44ec-aaf5-6aef86d24cdf","Type":"ContainerStarted","Data":"50272a6f57761cbe132b52ded8c8b616b0a374fd4d43bf61db1ff07e622423dc"} Feb 25 13:35:06 crc kubenswrapper[4815]: E0225 13:35:06.174384 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:06311600a491c689493552e7ff26e36df740fa4e7c143fca874bef19f24afb97\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-v825g" podUID="486a2442-b575-491d-9c18-4185e5a647fe" Feb 25 13:35:06 crc kubenswrapper[4815]: E0225 13:35:06.184027 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:a18f12497b7159b100fcfd72c7ba2273d0669a5c00600a9ff1333bca028f256a\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-8t7p2" podUID="16cb48d6-2014-44ec-aaf5-6aef86d24cdf" Feb 25 13:35:06 crc kubenswrapper[4815]: I0225 13:35:06.961868 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c43e580-fd0a-455b-92e9-c87d13dccccc" path="/var/lib/kubelet/pods/7c43e580-fd0a-455b-92e9-c87d13dccccc/volumes" Feb 25 13:35:07 crc kubenswrapper[4815]: I0225 13:35:07.146852 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert\") pod \"infra-operator-controller-manager-79d975b745-s5ztn\" (UID: \"9a554278-1139-4942-904a-ffc60adde6de\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" Feb 25 13:35:07 crc kubenswrapper[4815]: E0225 13:35:07.147035 4815 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 25 13:35:07 crc kubenswrapper[4815]: E0225 13:35:07.147086 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert podName:9a554278-1139-4942-904a-ffc60adde6de nodeName:}" failed. No retries permitted until 2026-02-25 13:35:11.147073208 +0000 UTC m=+868.948171262 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert") pod "infra-operator-controller-manager-79d975b745-s5ztn" (UID: "9a554278-1139-4942-904a-ffc60adde6de") : secret "infra-operator-webhook-server-cert" not found Feb 25 13:35:07 crc kubenswrapper[4815]: E0225 13:35:07.216367 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7l95p" podUID="22b232e5-5045-404a-a248-01babe4886c3" Feb 25 13:35:07 crc kubenswrapper[4815]: E0225 13:35:07.216942 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:f1158ec4d879c4646eee4323bc501eba4d377beb2ad6fbe08ed30070c441ac26\\\"\"" pod="openstack-operators/manila-operator-controller-manager-67d996989d-5fq96" podUID="4dc8d07d-3e17-40a8-ac19-187729a64c4d" Feb 25 13:35:07 crc kubenswrapper[4815]: E0225 13:35:07.216980 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:06311600a491c689493552e7ff26e36df740fa4e7c143fca874bef19f24afb97\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-v825g" podUID="486a2442-b575-491d-9c18-4185e5a647fe" Feb 25 13:35:07 crc kubenswrapper[4815]: E0225 13:35:07.217010 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:a18f12497b7159b100fcfd72c7ba2273d0669a5c00600a9ff1333bca028f256a\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-8t7p2" podUID="16cb48d6-2014-44ec-aaf5-6aef86d24cdf" Feb 25 13:35:07 crc kubenswrapper[4815]: E0225 13:35:07.217043 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:38e6a5bd24ab1684f22a64186fe99a7cdc7897eb7feb715ec1704eea7596dd98\\\"\"" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-grtsf" podUID="21de3543-3d6a-4adf-b104-f9fcaee4532b" Feb 25 13:35:07 crc kubenswrapper[4815]: E0225 13:35:07.218143 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f4143497c70c048a7733c284060347a0c74ef4e628aca22ee191e5bc9e4c7192\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-62tw8" podUID="697c602f-d6bc-478c-b89b-e3cc8aee4dcc" Feb 25 13:35:07 crc kubenswrapper[4815]: I0225 13:35:07.283808 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vt458" Feb 25 13:35:07 crc kubenswrapper[4815]: I0225 13:35:07.283855 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vt458" Feb 25 13:35:07 crc kubenswrapper[4815]: I0225 13:35:07.354057 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p\" (UID: \"43faa7ed-99a7-478c-8994-7a97333df231\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" Feb 25 13:35:07 crc kubenswrapper[4815]: E0225 13:35:07.354229 4815 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 13:35:07 crc kubenswrapper[4815]: E0225 13:35:07.354304 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert podName:43faa7ed-99a7-478c-8994-7a97333df231 nodeName:}" failed. No retries permitted until 2026-02-25 13:35:11.354276524 +0000 UTC m=+869.155374578 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" (UID: "43faa7ed-99a7-478c-8994-7a97333df231") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 13:35:07 crc kubenswrapper[4815]: I0225 13:35:07.765043 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:07 crc kubenswrapper[4815]: I0225 13:35:07.765088 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:07 crc kubenswrapper[4815]: E0225 13:35:07.765215 4815 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 25 13:35:07 crc kubenswrapper[4815]: E0225 13:35:07.765259 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs podName:509894d6-7814-4f85-9eed-f1946b029dbf nodeName:}" failed. No retries permitted until 2026-02-25 13:35:11.765246775 +0000 UTC m=+869.566344819 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs") pod "openstack-operator-controller-manager-5f47dfdb6c-slw4h" (UID: "509894d6-7814-4f85-9eed-f1946b029dbf") : secret "metrics-server-cert" not found Feb 25 13:35:07 crc kubenswrapper[4815]: E0225 13:35:07.765434 4815 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 25 13:35:07 crc kubenswrapper[4815]: E0225 13:35:07.765527 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs podName:509894d6-7814-4f85-9eed-f1946b029dbf nodeName:}" failed. No retries permitted until 2026-02-25 13:35:11.765483183 +0000 UTC m=+869.566581237 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs") pod "openstack-operator-controller-manager-5f47dfdb6c-slw4h" (UID: "509894d6-7814-4f85-9eed-f1946b029dbf") : secret "webhook-server-cert" not found Feb 25 13:35:08 crc kubenswrapper[4815]: I0225 13:35:08.507327 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vt458" podUID="2ff19dbc-fec6-46e8-b36a-e49dde7db51c" containerName="registry-server" probeResult="failure" output=< Feb 25 13:35:08 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Feb 25 13:35:08 crc kubenswrapper[4815]: > Feb 25 13:35:10 crc kubenswrapper[4815]: I0225 13:35:10.685858 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tddkm" Feb 25 13:35:10 crc kubenswrapper[4815]: I0225 13:35:10.685906 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tddkm" Feb 25 13:35:10 crc kubenswrapper[4815]: I0225 13:35:10.736231 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tddkm" Feb 25 13:35:11 crc kubenswrapper[4815]: I0225 13:35:11.228076 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert\") pod \"infra-operator-controller-manager-79d975b745-s5ztn\" (UID: \"9a554278-1139-4942-904a-ffc60adde6de\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" Feb 25 13:35:11 crc kubenswrapper[4815]: E0225 13:35:11.228292 4815 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 25 13:35:11 crc kubenswrapper[4815]: E0225 13:35:11.228357 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert podName:9a554278-1139-4942-904a-ffc60adde6de nodeName:}" failed. No retries permitted until 2026-02-25 13:35:19.228340741 +0000 UTC m=+877.029438795 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert") pod "infra-operator-controller-manager-79d975b745-s5ztn" (UID: "9a554278-1139-4942-904a-ffc60adde6de") : secret "infra-operator-webhook-server-cert" not found Feb 25 13:35:11 crc kubenswrapper[4815]: I0225 13:35:11.285461 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tddkm" Feb 25 13:35:11 crc kubenswrapper[4815]: I0225 13:35:11.323865 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tddkm"] Feb 25 13:35:11 crc kubenswrapper[4815]: I0225 13:35:11.430895 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p\" (UID: \"43faa7ed-99a7-478c-8994-7a97333df231\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" Feb 25 13:35:11 crc kubenswrapper[4815]: E0225 13:35:11.431059 4815 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 13:35:11 crc kubenswrapper[4815]: E0225 13:35:11.431125 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert podName:43faa7ed-99a7-478c-8994-7a97333df231 nodeName:}" failed. No retries permitted until 2026-02-25 13:35:19.431106978 +0000 UTC m=+877.232205032 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" (UID: "43faa7ed-99a7-478c-8994-7a97333df231") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 13:35:11 crc kubenswrapper[4815]: I0225 13:35:11.837346 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:11 crc kubenswrapper[4815]: I0225 13:35:11.837468 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:11 crc kubenswrapper[4815]: E0225 13:35:11.837666 4815 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 25 13:35:11 crc kubenswrapper[4815]: E0225 13:35:11.837703 4815 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 25 13:35:11 crc kubenswrapper[4815]: E0225 13:35:11.837770 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs podName:509894d6-7814-4f85-9eed-f1946b029dbf nodeName:}" failed. No retries permitted until 2026-02-25 13:35:19.837741125 +0000 UTC m=+877.638839219 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs") pod "openstack-operator-controller-manager-5f47dfdb6c-slw4h" (UID: "509894d6-7814-4f85-9eed-f1946b029dbf") : secret "webhook-server-cert" not found Feb 25 13:35:11 crc kubenswrapper[4815]: E0225 13:35:11.837799 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs podName:509894d6-7814-4f85-9eed-f1946b029dbf nodeName:}" failed. No retries permitted until 2026-02-25 13:35:19.837786757 +0000 UTC m=+877.638884851 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs") pod "openstack-operator-controller-manager-5f47dfdb6c-slw4h" (UID: "509894d6-7814-4f85-9eed-f1946b029dbf") : secret "metrics-server-cert" not found Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.262731 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tddkm" podUID="59dc6b8b-e3e2-4f25-91bf-310d17bbd603" containerName="registry-server" containerID="cri-o://d419ac199f099c4ac254ef71a5cdaa1d318d0547c94f779cd8aa72515ebb27ad" gracePeriod=2 Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.395868 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-48hkj"] Feb 25 13:35:13 crc kubenswrapper[4815]: E0225 13:35:13.396221 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c43e580-fd0a-455b-92e9-c87d13dccccc" containerName="extract-content" Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.396235 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c43e580-fd0a-455b-92e9-c87d13dccccc" containerName="extract-content" Feb 25 13:35:13 crc kubenswrapper[4815]: E0225 13:35:13.396264 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c43e580-fd0a-455b-92e9-c87d13dccccc" containerName="registry-server" Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.396272 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c43e580-fd0a-455b-92e9-c87d13dccccc" containerName="registry-server" Feb 25 13:35:13 crc kubenswrapper[4815]: E0225 13:35:13.396281 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c43e580-fd0a-455b-92e9-c87d13dccccc" containerName="extract-utilities" Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.396288 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c43e580-fd0a-455b-92e9-c87d13dccccc" containerName="extract-utilities" Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.396442 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c43e580-fd0a-455b-92e9-c87d13dccccc" containerName="registry-server" Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.397494 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-48hkj" Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.400468 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-48hkj"] Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.463852 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-utilities\") pod \"community-operators-48hkj\" (UID: \"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4\") " pod="openshift-marketplace/community-operators-48hkj" Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.463898 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-catalog-content\") pod \"community-operators-48hkj\" (UID: \"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4\") " pod="openshift-marketplace/community-operators-48hkj" Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.464132 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpxlb\" (UniqueName: \"kubernetes.io/projected/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-kube-api-access-gpxlb\") pod \"community-operators-48hkj\" (UID: \"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4\") " pod="openshift-marketplace/community-operators-48hkj" Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.565494 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-utilities\") pod \"community-operators-48hkj\" (UID: \"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4\") " pod="openshift-marketplace/community-operators-48hkj" Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.565555 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-catalog-content\") pod \"community-operators-48hkj\" (UID: \"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4\") " pod="openshift-marketplace/community-operators-48hkj" Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.565606 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpxlb\" (UniqueName: \"kubernetes.io/projected/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-kube-api-access-gpxlb\") pod \"community-operators-48hkj\" (UID: \"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4\") " pod="openshift-marketplace/community-operators-48hkj" Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.566052 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-catalog-content\") pod \"community-operators-48hkj\" (UID: \"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4\") " pod="openshift-marketplace/community-operators-48hkj" Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.566115 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-utilities\") pod \"community-operators-48hkj\" (UID: \"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4\") " pod="openshift-marketplace/community-operators-48hkj" Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.587554 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpxlb\" (UniqueName: \"kubernetes.io/projected/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-kube-api-access-gpxlb\") pod \"community-operators-48hkj\" (UID: \"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4\") " pod="openshift-marketplace/community-operators-48hkj" Feb 25 13:35:13 crc kubenswrapper[4815]: I0225 13:35:13.720387 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-48hkj" Feb 25 13:35:14 crc kubenswrapper[4815]: I0225 13:35:14.273848 4815 generic.go:334] "Generic (PLEG): container finished" podID="59dc6b8b-e3e2-4f25-91bf-310d17bbd603" containerID="d419ac199f099c4ac254ef71a5cdaa1d318d0547c94f779cd8aa72515ebb27ad" exitCode=0 Feb 25 13:35:14 crc kubenswrapper[4815]: I0225 13:35:14.273889 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tddkm" event={"ID":"59dc6b8b-e3e2-4f25-91bf-310d17bbd603","Type":"ContainerDied","Data":"d419ac199f099c4ac254ef71a5cdaa1d318d0547c94f779cd8aa72515ebb27ad"} Feb 25 13:35:17 crc kubenswrapper[4815]: I0225 13:35:17.346639 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vt458" Feb 25 13:35:17 crc kubenswrapper[4815]: I0225 13:35:17.398021 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vt458" Feb 25 13:35:18 crc kubenswrapper[4815]: E0225 13:35:18.379704 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:90ad8fd8c1889b6be77925016532218eb6149d2c1c8535a5f9f1775c776fa6cc" Feb 25 13:35:18 crc kubenswrapper[4815]: E0225 13:35:18.380405 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:90ad8fd8c1889b6be77925016532218eb6149d2c1c8535a5f9f1775c776fa6cc,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mqbbs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-868647ff47-zwknl_openstack-operators(62c00caf-05cf-4229-9349-341a2b597b12): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 13:35:18 crc kubenswrapper[4815]: E0225 13:35:18.381862 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-zwknl" podUID="62c00caf-05cf-4229-9349-341a2b597b12" Feb 25 13:35:18 crc kubenswrapper[4815]: I0225 13:35:18.572589 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vt458"] Feb 25 13:35:19 crc kubenswrapper[4815]: E0225 13:35:19.048023 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:4eb8fab5530a08915d3ab3e11e2808aeae16c8a220ed34ee04a186b2ae2303dc" Feb 25 13:35:19 crc kubenswrapper[4815]: E0225 13:35:19.048277 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:4eb8fab5530a08915d3ab3e11e2808aeae16c8a220ed34ee04a186b2ae2303dc,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w8vc6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-589c568786-rwzjh_openstack-operators(136c32bc-0a85-4a36-b527-a787fa7158f1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 13:35:19 crc kubenswrapper[4815]: E0225 13:35:19.049745 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-rwzjh" podUID="136c32bc-0a85-4a36-b527-a787fa7158f1" Feb 25 13:35:19 crc kubenswrapper[4815]: I0225 13:35:19.261948 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert\") pod \"infra-operator-controller-manager-79d975b745-s5ztn\" (UID: \"9a554278-1139-4942-904a-ffc60adde6de\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" Feb 25 13:35:19 crc kubenswrapper[4815]: I0225 13:35:19.268379 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9a554278-1139-4942-904a-ffc60adde6de-cert\") pod \"infra-operator-controller-manager-79d975b745-s5ztn\" (UID: \"9a554278-1139-4942-904a-ffc60adde6de\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" Feb 25 13:35:19 crc kubenswrapper[4815]: I0225 13:35:19.309061 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vt458" podUID="2ff19dbc-fec6-46e8-b36a-e49dde7db51c" containerName="registry-server" containerID="cri-o://949c8fc460788e63ce64de8cdc808e03544a91c144fe4133a9f376e186ac5332" gracePeriod=2 Feb 25 13:35:19 crc kubenswrapper[4815]: E0225 13:35:19.312561 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:90ad8fd8c1889b6be77925016532218eb6149d2c1c8535a5f9f1775c776fa6cc\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-zwknl" podUID="62c00caf-05cf-4229-9349-341a2b597b12" Feb 25 13:35:19 crc kubenswrapper[4815]: E0225 13:35:19.312668 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:4eb8fab5530a08915d3ab3e11e2808aeae16c8a220ed34ee04a186b2ae2303dc\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-rwzjh" podUID="136c32bc-0a85-4a36-b527-a787fa7158f1" Feb 25 13:35:19 crc kubenswrapper[4815]: I0225 13:35:19.343900 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" Feb 25 13:35:19 crc kubenswrapper[4815]: I0225 13:35:19.466367 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p\" (UID: \"43faa7ed-99a7-478c-8994-7a97333df231\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" Feb 25 13:35:19 crc kubenswrapper[4815]: E0225 13:35:19.466646 4815 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 13:35:19 crc kubenswrapper[4815]: E0225 13:35:19.466742 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert podName:43faa7ed-99a7-478c-8994-7a97333df231 nodeName:}" failed. No retries permitted until 2026-02-25 13:35:35.466715754 +0000 UTC m=+893.267813838 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" (UID: "43faa7ed-99a7-478c-8994-7a97333df231") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 13:35:19 crc kubenswrapper[4815]: E0225 13:35:19.587165 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:14ae1fb8d065e2317959ce7490a878dc87731d27ebf40259f801ba1a83cfefcf" Feb 25 13:35:19 crc kubenswrapper[4815]: E0225 13:35:19.587417 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:14ae1fb8d065e2317959ce7490a878dc87731d27ebf40259f801ba1a83cfefcf,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q5mz2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-6bd4687957-vnb57_openstack-operators(6bab4cd6-5eff-409c-83e3-ad5c1a6b6444): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 13:35:19 crc kubenswrapper[4815]: E0225 13:35:19.588759 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-vnb57" podUID="6bab4cd6-5eff-409c-83e3-ad5c1a6b6444" Feb 25 13:35:19 crc kubenswrapper[4815]: I0225 13:35:19.871612 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:19 crc kubenswrapper[4815]: I0225 13:35:19.871816 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:19 crc kubenswrapper[4815]: E0225 13:35:19.871851 4815 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 25 13:35:19 crc kubenswrapper[4815]: E0225 13:35:19.871934 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs podName:509894d6-7814-4f85-9eed-f1946b029dbf nodeName:}" failed. No retries permitted until 2026-02-25 13:35:35.871916998 +0000 UTC m=+893.673015052 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs") pod "openstack-operator-controller-manager-5f47dfdb6c-slw4h" (UID: "509894d6-7814-4f85-9eed-f1946b029dbf") : secret "metrics-server-cert" not found Feb 25 13:35:19 crc kubenswrapper[4815]: E0225 13:35:19.871959 4815 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 25 13:35:19 crc kubenswrapper[4815]: E0225 13:35:19.872021 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs podName:509894d6-7814-4f85-9eed-f1946b029dbf nodeName:}" failed. No retries permitted until 2026-02-25 13:35:35.87200029 +0000 UTC m=+893.673098404 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs") pod "openstack-operator-controller-manager-5f47dfdb6c-slw4h" (UID: "509894d6-7814-4f85-9eed-f1946b029dbf") : secret "webhook-server-cert" not found Feb 25 13:35:20 crc kubenswrapper[4815]: I0225 13:35:20.317797 4815 generic.go:334] "Generic (PLEG): container finished" podID="2ff19dbc-fec6-46e8-b36a-e49dde7db51c" containerID="949c8fc460788e63ce64de8cdc808e03544a91c144fe4133a9f376e186ac5332" exitCode=0 Feb 25 13:35:20 crc kubenswrapper[4815]: I0225 13:35:20.317882 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vt458" event={"ID":"2ff19dbc-fec6-46e8-b36a-e49dde7db51c","Type":"ContainerDied","Data":"949c8fc460788e63ce64de8cdc808e03544a91c144fe4133a9f376e186ac5332"} Feb 25 13:35:20 crc kubenswrapper[4815]: E0225 13:35:20.319565 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:14ae1fb8d065e2317959ce7490a878dc87731d27ebf40259f801ba1a83cfefcf\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-vnb57" podUID="6bab4cd6-5eff-409c-83e3-ad5c1a6b6444" Feb 25 13:35:20 crc kubenswrapper[4815]: E0225 13:35:20.389696 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04" Feb 25 13:35:20 crc kubenswrapper[4815]: E0225 13:35:20.390278 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q9szd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-68f46476f-8vr5p_openstack-operators(0673a594-2080-4325-80ce-f5597ad337c7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 13:35:20 crc kubenswrapper[4815]: E0225 13:35:20.391478 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-8vr5p" podUID="0673a594-2080-4325-80ce-f5597ad337c7" Feb 25 13:35:20 crc kubenswrapper[4815]: E0225 13:35:20.686905 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d419ac199f099c4ac254ef71a5cdaa1d318d0547c94f779cd8aa72515ebb27ad is running failed: container process not found" containerID="d419ac199f099c4ac254ef71a5cdaa1d318d0547c94f779cd8aa72515ebb27ad" cmd=["grpc_health_probe","-addr=:50051"] Feb 25 13:35:20 crc kubenswrapper[4815]: E0225 13:35:20.687433 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d419ac199f099c4ac254ef71a5cdaa1d318d0547c94f779cd8aa72515ebb27ad is running failed: container process not found" containerID="d419ac199f099c4ac254ef71a5cdaa1d318d0547c94f779cd8aa72515ebb27ad" cmd=["grpc_health_probe","-addr=:50051"] Feb 25 13:35:20 crc kubenswrapper[4815]: E0225 13:35:20.688199 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d419ac199f099c4ac254ef71a5cdaa1d318d0547c94f779cd8aa72515ebb27ad is running failed: container process not found" containerID="d419ac199f099c4ac254ef71a5cdaa1d318d0547c94f779cd8aa72515ebb27ad" cmd=["grpc_health_probe","-addr=:50051"] Feb 25 13:35:20 crc kubenswrapper[4815]: E0225 13:35:20.688232 4815 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d419ac199f099c4ac254ef71a5cdaa1d318d0547c94f779cd8aa72515ebb27ad is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-tddkm" podUID="59dc6b8b-e3e2-4f25-91bf-310d17bbd603" containerName="registry-server" Feb 25 13:35:20 crc kubenswrapper[4815]: E0225 13:35:20.957546 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838" Feb 25 13:35:20 crc kubenswrapper[4815]: E0225 13:35:20.957742 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-96nkl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-567668f5cf-z4jrr_openstack-operators(b30720e2-30f0-480e-a44f-2b409be766c6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 13:35:20 crc kubenswrapper[4815]: E0225 13:35:20.958921 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-z4jrr" podUID="b30720e2-30f0-480e-a44f-2b409be766c6" Feb 25 13:35:21 crc kubenswrapper[4815]: E0225 13:35:21.328797 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838\\\"\"" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-z4jrr" podUID="b30720e2-30f0-480e-a44f-2b409be766c6" Feb 25 13:35:21 crc kubenswrapper[4815]: E0225 13:35:21.329065 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-8vr5p" podUID="0673a594-2080-4325-80ce-f5597ad337c7" Feb 25 13:35:21 crc kubenswrapper[4815]: E0225 13:35:21.567793 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1" Feb 25 13:35:21 crc kubenswrapper[4815]: E0225 13:35:21.568636 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xqqrr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-b4d948c87-rx4fg_openstack-operators(c64a07be-3cda-473e-89b6-fd0347519c93): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 13:35:21 crc kubenswrapper[4815]: E0225 13:35:21.569834 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-rx4fg" podUID="c64a07be-3cda-473e-89b6-fd0347519c93" Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.633485 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vt458" Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.643372 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tddkm" Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.803466 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-utilities\") pod \"2ff19dbc-fec6-46e8-b36a-e49dde7db51c\" (UID: \"2ff19dbc-fec6-46e8-b36a-e49dde7db51c\") " Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.803579 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-utilities\") pod \"59dc6b8b-e3e2-4f25-91bf-310d17bbd603\" (UID: \"59dc6b8b-e3e2-4f25-91bf-310d17bbd603\") " Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.803657 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2tjh\" (UniqueName: \"kubernetes.io/projected/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-kube-api-access-k2tjh\") pod \"2ff19dbc-fec6-46e8-b36a-e49dde7db51c\" (UID: \"2ff19dbc-fec6-46e8-b36a-e49dde7db51c\") " Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.803707 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-catalog-content\") pod \"2ff19dbc-fec6-46e8-b36a-e49dde7db51c\" (UID: \"2ff19dbc-fec6-46e8-b36a-e49dde7db51c\") " Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.803726 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgwvt\" (UniqueName: \"kubernetes.io/projected/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-kube-api-access-cgwvt\") pod \"59dc6b8b-e3e2-4f25-91bf-310d17bbd603\" (UID: \"59dc6b8b-e3e2-4f25-91bf-310d17bbd603\") " Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.803746 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-catalog-content\") pod \"59dc6b8b-e3e2-4f25-91bf-310d17bbd603\" (UID: \"59dc6b8b-e3e2-4f25-91bf-310d17bbd603\") " Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.804327 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-utilities" (OuterVolumeSpecName: "utilities") pod "2ff19dbc-fec6-46e8-b36a-e49dde7db51c" (UID: "2ff19dbc-fec6-46e8-b36a-e49dde7db51c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.809628 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-kube-api-access-cgwvt" (OuterVolumeSpecName: "kube-api-access-cgwvt") pod "59dc6b8b-e3e2-4f25-91bf-310d17bbd603" (UID: "59dc6b8b-e3e2-4f25-91bf-310d17bbd603"). InnerVolumeSpecName "kube-api-access-cgwvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.809914 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-kube-api-access-k2tjh" (OuterVolumeSpecName: "kube-api-access-k2tjh") pod "2ff19dbc-fec6-46e8-b36a-e49dde7db51c" (UID: "2ff19dbc-fec6-46e8-b36a-e49dde7db51c"). InnerVolumeSpecName "kube-api-access-k2tjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.811168 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-utilities" (OuterVolumeSpecName: "utilities") pod "59dc6b8b-e3e2-4f25-91bf-310d17bbd603" (UID: "59dc6b8b-e3e2-4f25-91bf-310d17bbd603"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.853430 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59dc6b8b-e3e2-4f25-91bf-310d17bbd603" (UID: "59dc6b8b-e3e2-4f25-91bf-310d17bbd603"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.905666 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.905703 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.905717 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2tjh\" (UniqueName: \"kubernetes.io/projected/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-kube-api-access-k2tjh\") on node \"crc\" DevicePath \"\"" Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.905731 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgwvt\" (UniqueName: \"kubernetes.io/projected/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-kube-api-access-cgwvt\") on node \"crc\" DevicePath \"\"" Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.905745 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59dc6b8b-e3e2-4f25-91bf-310d17bbd603-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:35:21 crc kubenswrapper[4815]: I0225 13:35:21.932068 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ff19dbc-fec6-46e8-b36a-e49dde7db51c" (UID: "2ff19dbc-fec6-46e8-b36a-e49dde7db51c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:35:22 crc kubenswrapper[4815]: I0225 13:35:22.007483 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ff19dbc-fec6-46e8-b36a-e49dde7db51c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:35:22 crc kubenswrapper[4815]: I0225 13:35:22.338463 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tddkm" event={"ID":"59dc6b8b-e3e2-4f25-91bf-310d17bbd603","Type":"ContainerDied","Data":"9c58fd8b301072c4cf0acf81b294316b94bbd34bd73e480cbe757be9387ab4e3"} Feb 25 13:35:22 crc kubenswrapper[4815]: I0225 13:35:22.338522 4815 scope.go:117] "RemoveContainer" containerID="d419ac199f099c4ac254ef71a5cdaa1d318d0547c94f779cd8aa72515ebb27ad" Feb 25 13:35:22 crc kubenswrapper[4815]: I0225 13:35:22.338757 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tddkm" Feb 25 13:35:22 crc kubenswrapper[4815]: I0225 13:35:22.341369 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vt458" Feb 25 13:35:22 crc kubenswrapper[4815]: I0225 13:35:22.341604 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vt458" event={"ID":"2ff19dbc-fec6-46e8-b36a-e49dde7db51c","Type":"ContainerDied","Data":"2c528fb19f9c716e34dec868c18069058aabce3c8644457b55ee4a8aeef8ee77"} Feb 25 13:35:22 crc kubenswrapper[4815]: E0225 13:35:22.342070 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-rx4fg" podUID="c64a07be-3cda-473e-89b6-fd0347519c93" Feb 25 13:35:22 crc kubenswrapper[4815]: I0225 13:35:22.375136 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tddkm"] Feb 25 13:35:22 crc kubenswrapper[4815]: I0225 13:35:22.382970 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tddkm"] Feb 25 13:35:22 crc kubenswrapper[4815]: I0225 13:35:22.391727 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vt458"] Feb 25 13:35:22 crc kubenswrapper[4815]: I0225 13:35:22.397750 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vt458"] Feb 25 13:35:22 crc kubenswrapper[4815]: I0225 13:35:22.958881 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ff19dbc-fec6-46e8-b36a-e49dde7db51c" path="/var/lib/kubelet/pods/2ff19dbc-fec6-46e8-b36a-e49dde7db51c/volumes" Feb 25 13:35:22 crc kubenswrapper[4815]: I0225 13:35:22.960089 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59dc6b8b-e3e2-4f25-91bf-310d17bbd603" path="/var/lib/kubelet/pods/59dc6b8b-e3e2-4f25-91bf-310d17bbd603/volumes" Feb 25 13:35:23 crc kubenswrapper[4815]: I0225 13:35:23.245566 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn"] Feb 25 13:35:23 crc kubenswrapper[4815]: I0225 13:35:23.274940 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-48hkj"] Feb 25 13:35:23 crc kubenswrapper[4815]: W0225 13:35:23.643626 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a554278_1139_4942_904a_ffc60adde6de.slice/crio-528b0888bd6300618163a07d29e84b78be8ab67fbc5f2f8f8113910d481a5803 WatchSource:0}: Error finding container 528b0888bd6300618163a07d29e84b78be8ab67fbc5f2f8f8113910d481a5803: Status 404 returned error can't find the container with id 528b0888bd6300618163a07d29e84b78be8ab67fbc5f2f8f8113910d481a5803 Feb 25 13:35:23 crc kubenswrapper[4815]: I0225 13:35:23.654479 4815 scope.go:117] "RemoveContainer" containerID="e687322c626dafb096300a9c8ba7e52208ef18d15312da9bcd8afd9d613c2ff3" Feb 25 13:35:24 crc kubenswrapper[4815]: I0225 13:35:24.370866 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-48hkj" event={"ID":"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4","Type":"ContainerStarted","Data":"3b8d8d90fe0de3ebbd4ca814ca65569cc347b957e67a804e5cc0046ffe228d50"} Feb 25 13:35:24 crc kubenswrapper[4815]: I0225 13:35:24.372260 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" event={"ID":"9a554278-1139-4942-904a-ffc60adde6de","Type":"ContainerStarted","Data":"528b0888bd6300618163a07d29e84b78be8ab67fbc5f2f8f8113910d481a5803"} Feb 25 13:35:25 crc kubenswrapper[4815]: I0225 13:35:25.385180 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-h9n4d" event={"ID":"bace17a8-ce7d-4df9-8259-68931eeecde8","Type":"ContainerStarted","Data":"9d455771dce0b8b56eaf9d4b79489650bd3b966773351d3e7956b327ee3ba5aa"} Feb 25 13:35:25 crc kubenswrapper[4815]: I0225 13:35:25.385714 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-h9n4d" Feb 25 13:35:25 crc kubenswrapper[4815]: I0225 13:35:25.405690 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-h9n4d" podStartSLOduration=5.330637999 podStartE2EDuration="22.405670259s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:04.473035962 +0000 UTC m=+862.274134016" lastFinishedPulling="2026-02-25 13:35:21.548068212 +0000 UTC m=+879.349166276" observedRunningTime="2026-02-25 13:35:25.399139847 +0000 UTC m=+883.200237901" watchObservedRunningTime="2026-02-25 13:35:25.405670259 +0000 UTC m=+883.206768333" Feb 25 13:35:27 crc kubenswrapper[4815]: I0225 13:35:27.384879 4815 scope.go:117] "RemoveContainer" containerID="91ae3ac326d06086bce047b187279b1041e096036daa45a559cb9c58cb9415e9" Feb 25 13:35:28 crc kubenswrapper[4815]: I0225 13:35:28.407946 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9qrdg" event={"ID":"3b1a8f42-91a7-46b3-96c6-0ef20a126678","Type":"ContainerStarted","Data":"4ed6e3451cec3f368d613d44f628de61274bc9a6e9e5b380e999077b00a4147e"} Feb 25 13:35:28 crc kubenswrapper[4815]: I0225 13:35:28.408891 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9qrdg" Feb 25 13:35:28 crc kubenswrapper[4815]: I0225 13:35:28.411592 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-gw8z5" event={"ID":"9fa8caf9-fab3-44d2-9841-593558e9a690","Type":"ContainerStarted","Data":"e08f5bede14bd2d5796ae24fe68c83803193de4fa54e3378ca0930d82bdcf9cc"} Feb 25 13:35:28 crc kubenswrapper[4815]: I0225 13:35:28.411838 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-gw8z5" Feb 25 13:35:28 crc kubenswrapper[4815]: I0225 13:35:28.417096 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-qm5sw" event={"ID":"c4956173-674d-4b10-9a3f-9996f9190299","Type":"ContainerStarted","Data":"62b05b2cb8199224e9edae695e66f16aeb26fa2b9327f8ad2f4933bdbc5ca422"} Feb 25 13:35:28 crc kubenswrapper[4815]: I0225 13:35:28.417240 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-qm5sw" Feb 25 13:35:28 crc kubenswrapper[4815]: I0225 13:35:28.428777 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9qrdg" podStartSLOduration=9.086317396 podStartE2EDuration="25.428759133s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:05.193632074 +0000 UTC m=+862.994730128" lastFinishedPulling="2026-02-25 13:35:21.536073811 +0000 UTC m=+879.337171865" observedRunningTime="2026-02-25 13:35:28.423413148 +0000 UTC m=+886.224511202" watchObservedRunningTime="2026-02-25 13:35:28.428759133 +0000 UTC m=+886.229857187" Feb 25 13:35:28 crc kubenswrapper[4815]: I0225 13:35:28.442671 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-qm5sw" podStartSLOduration=8.375265327 podStartE2EDuration="25.442656382s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:04.480673607 +0000 UTC m=+862.281771661" lastFinishedPulling="2026-02-25 13:35:21.548064662 +0000 UTC m=+879.349162716" observedRunningTime="2026-02-25 13:35:28.440594259 +0000 UTC m=+886.241692313" watchObservedRunningTime="2026-02-25 13:35:28.442656382 +0000 UTC m=+886.243754436" Feb 25 13:35:28 crc kubenswrapper[4815]: I0225 13:35:28.458748 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-gw8z5" podStartSLOduration=8.595447183 podStartE2EDuration="25.458728759s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:04.674039694 +0000 UTC m=+862.475137748" lastFinishedPulling="2026-02-25 13:35:21.53732127 +0000 UTC m=+879.338419324" observedRunningTime="2026-02-25 13:35:28.454528389 +0000 UTC m=+886.255626463" watchObservedRunningTime="2026-02-25 13:35:28.458728759 +0000 UTC m=+886.259826823" Feb 25 13:35:29 crc kubenswrapper[4815]: I0225 13:35:29.502947 4815 scope.go:117] "RemoveContainer" containerID="949c8fc460788e63ce64de8cdc808e03544a91c144fe4133a9f376e186ac5332" Feb 25 13:35:29 crc kubenswrapper[4815]: I0225 13:35:29.568745 4815 scope.go:117] "RemoveContainer" containerID="dc4436fb78751730fc98b1343e2e2fee687e3dc621d385adbb433dfc094401f4" Feb 25 13:35:29 crc kubenswrapper[4815]: I0225 13:35:29.667525 4815 scope.go:117] "RemoveContainer" containerID="bfb0aed0c74c3a59dab1f49deec4a2b6ea1beea423b2d384678ea0ce6ccbe7ab" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.430127 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-8t7p2" event={"ID":"16cb48d6-2014-44ec-aaf5-6aef86d24cdf","Type":"ContainerStarted","Data":"0381e6541c84d820a9a5ac15f2a0dc6de1431dc5de2f0fdf649b5a40a5a01200"} Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.430346 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-8t7p2" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.430992 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-ccl5w" event={"ID":"33e59af6-b9bb-4b55-b6a9-c6963e269b94","Type":"ContainerStarted","Data":"9bea126cc2f5d609f0597fa75e745a69c62d4500892207bf98e4766ddc36e6e0"} Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.431127 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-ccl5w" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.432325 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-52psd" event={"ID":"12b1d896-ab7e-4097-afa4-791d39e1924e","Type":"ContainerStarted","Data":"c973860d71947848db44a0dd2acbde19b364f8c35e10e0b63b13956d142926b3"} Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.432530 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-52psd" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.433913 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7l95p" event={"ID":"22b232e5-5045-404a-a248-01babe4886c3","Type":"ContainerStarted","Data":"fa07f4638362335d403878388528fd3d7b5837ff26b7de3e754a743a5d8f7d98"} Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.435128 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-v825g" event={"ID":"486a2442-b575-491d-9c18-4185e5a647fe","Type":"ContainerStarted","Data":"0e03f553db56f55949805881acd7ddd450f3a138a12e55fe8a0f229abfdc0f82"} Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.435370 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-v825g" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.436487 4815 generic.go:334] "Generic (PLEG): container finished" podID="eec5ea92-0b0f-4c61-a0ee-074a5a21eba4" containerID="863f30e473201d0c095c68514d1f388d7d6877e98546f2ed75acb86dbf6b17af" exitCode=0 Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.436545 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-48hkj" event={"ID":"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4","Type":"ContainerDied","Data":"863f30e473201d0c095c68514d1f388d7d6877e98546f2ed75acb86dbf6b17af"} Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.438695 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" event={"ID":"9a554278-1139-4942-904a-ffc60adde6de","Type":"ContainerStarted","Data":"4895a81270ff544064329d420703a4d2dc6b4bc1dbdcca845e4069df80ac68f2"} Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.440416 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.471517 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-8t7p2" podStartSLOduration=3.146580135 podStartE2EDuration="27.471485967s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:05.219948786 +0000 UTC m=+863.021046840" lastFinishedPulling="2026-02-25 13:35:29.544854618 +0000 UTC m=+887.345952672" observedRunningTime="2026-02-25 13:35:30.458555058 +0000 UTC m=+888.259653112" watchObservedRunningTime="2026-02-25 13:35:30.471485967 +0000 UTC m=+888.272584021" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.479198 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-qtr8c" event={"ID":"b7cddc29-1c58-4edd-992f-165386b751ba","Type":"ContainerStarted","Data":"84f98b3f095e2f514211ff8f7ba231aa1010e40aca6fa5e297b1770b260e0483"} Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.480215 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-qtr8c" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.501263 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-ccl5w" podStartSLOduration=10.695442688 podStartE2EDuration="27.501241197s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:04.754959285 +0000 UTC m=+862.556057339" lastFinishedPulling="2026-02-25 13:35:21.560757784 +0000 UTC m=+879.361855848" observedRunningTime="2026-02-25 13:35:30.497537692 +0000 UTC m=+888.298635756" watchObservedRunningTime="2026-02-25 13:35:30.501241197 +0000 UTC m=+888.302339251" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.501567 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-62tw8" event={"ID":"697c602f-d6bc-478c-b89b-e3cc8aee4dcc","Type":"ContainerStarted","Data":"506c189057e07978adf2084ffd7a932a8dc805ed6665dd055762df3612fd57e3"} Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.501806 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-62tw8" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.514962 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-csfk2" event={"ID":"50963101-0a9e-453d-925b-ac6f08c545c4","Type":"ContainerStarted","Data":"b93373410327f84e2fe6c3d8824699f3a6089da44b375a9e72efdd606202bd79"} Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.515601 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-csfk2" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.520012 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" podStartSLOduration=21.668519095 podStartE2EDuration="27.519990796s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:23.681466969 +0000 UTC m=+881.482565023" lastFinishedPulling="2026-02-25 13:35:29.53293867 +0000 UTC m=+887.334036724" observedRunningTime="2026-02-25 13:35:30.517654304 +0000 UTC m=+888.318752358" watchObservedRunningTime="2026-02-25 13:35:30.519990796 +0000 UTC m=+888.321088850" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.523082 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-grtsf" event={"ID":"21de3543-3d6a-4adf-b104-f9fcaee4532b","Type":"ContainerStarted","Data":"9d4124f3b7620b575aae70a4160686e1ebab8a51b1d0a86bafac9832cb48db26"} Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.523300 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-grtsf" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.524664 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-67d996989d-5fq96" event={"ID":"4dc8d07d-3e17-40a8-ac19-187729a64c4d","Type":"ContainerStarted","Data":"5380beb56281dfe996181270a8003d17ca4220e5503d41f5c8f4323d262852d6"} Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.524888 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-67d996989d-5fq96" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.599185 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7l95p" podStartSLOduration=3.301215215 podStartE2EDuration="27.599169094s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:05.233958359 +0000 UTC m=+863.035056413" lastFinishedPulling="2026-02-25 13:35:29.531912238 +0000 UTC m=+887.333010292" observedRunningTime="2026-02-25 13:35:30.597704548 +0000 UTC m=+888.398802602" watchObservedRunningTime="2026-02-25 13:35:30.599169094 +0000 UTC m=+888.400267148" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.600373 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-52psd" podStartSLOduration=10.496856599000001 podStartE2EDuration="27.60036775s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:05.116925392 +0000 UTC m=+862.918023446" lastFinishedPulling="2026-02-25 13:35:22.220436543 +0000 UTC m=+880.021534597" observedRunningTime="2026-02-25 13:35:30.582729946 +0000 UTC m=+888.383828000" watchObservedRunningTime="2026-02-25 13:35:30.60036775 +0000 UTC m=+888.401465804" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.619459 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-v825g" podStartSLOduration=8.612368777 podStartE2EDuration="27.619445441s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:05.233959869 +0000 UTC m=+863.035057923" lastFinishedPulling="2026-02-25 13:35:24.241036543 +0000 UTC m=+882.042134587" observedRunningTime="2026-02-25 13:35:30.616368036 +0000 UTC m=+888.417466090" watchObservedRunningTime="2026-02-25 13:35:30.619445441 +0000 UTC m=+888.420543495" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.637084 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-qtr8c" podStartSLOduration=11.264435373 podStartE2EDuration="27.637067675s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:05.163364037 +0000 UTC m=+862.964462091" lastFinishedPulling="2026-02-25 13:35:21.535996339 +0000 UTC m=+879.337094393" observedRunningTime="2026-02-25 13:35:30.635200607 +0000 UTC m=+888.436298661" watchObservedRunningTime="2026-02-25 13:35:30.637067675 +0000 UTC m=+888.438165729" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.656572 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-62tw8" podStartSLOduration=4.751767677 podStartE2EDuration="27.656549767s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:05.209403551 +0000 UTC m=+863.010501605" lastFinishedPulling="2026-02-25 13:35:28.114185641 +0000 UTC m=+885.915283695" observedRunningTime="2026-02-25 13:35:30.651958045 +0000 UTC m=+888.453056089" watchObservedRunningTime="2026-02-25 13:35:30.656549767 +0000 UTC m=+888.457647821" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.697274 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-67d996989d-5fq96" podStartSLOduration=5.523155309 podStartE2EDuration="27.697258985s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:05.210619769 +0000 UTC m=+863.011717823" lastFinishedPulling="2026-02-25 13:35:27.384723435 +0000 UTC m=+885.185821499" observedRunningTime="2026-02-25 13:35:30.694678036 +0000 UTC m=+888.495776090" watchObservedRunningTime="2026-02-25 13:35:30.697258985 +0000 UTC m=+888.498357039" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.712852 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-grtsf" podStartSLOduration=8.706373641999999 podStartE2EDuration="27.712836787s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:05.233957139 +0000 UTC m=+863.035055193" lastFinishedPulling="2026-02-25 13:35:24.240420284 +0000 UTC m=+882.041518338" observedRunningTime="2026-02-25 13:35:30.707662587 +0000 UTC m=+888.508760641" watchObservedRunningTime="2026-02-25 13:35:30.712836787 +0000 UTC m=+888.513934841" Feb 25 13:35:30 crc kubenswrapper[4815]: I0225 13:35:30.722894 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-csfk2" podStartSLOduration=10.896609204 podStartE2EDuration="27.722877767s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:04.721842621 +0000 UTC m=+862.522940675" lastFinishedPulling="2026-02-25 13:35:21.548111194 +0000 UTC m=+879.349209238" observedRunningTime="2026-02-25 13:35:30.718689557 +0000 UTC m=+888.519787611" watchObservedRunningTime="2026-02-25 13:35:30.722877767 +0000 UTC m=+888.523975821" Feb 25 13:35:33 crc kubenswrapper[4815]: I0225 13:35:33.511632 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-h9n4d" Feb 25 13:35:33 crc kubenswrapper[4815]: I0225 13:35:33.549386 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-qm5sw" Feb 25 13:35:33 crc kubenswrapper[4815]: I0225 13:35:33.557640 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-gw8z5" Feb 25 13:35:33 crc kubenswrapper[4815]: I0225 13:35:33.978132 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9qrdg" Feb 25 13:35:34 crc kubenswrapper[4815]: I0225 13:35:34.050027 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-8t7p2" Feb 25 13:35:34 crc kubenswrapper[4815]: I0225 13:35:34.055699 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-qtr8c" Feb 25 13:35:34 crc kubenswrapper[4815]: I0225 13:35:34.419380 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-grtsf" Feb 25 13:35:34 crc kubenswrapper[4815]: I0225 13:35:34.475992 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-v825g" Feb 25 13:35:35 crc kubenswrapper[4815]: I0225 13:35:35.510785 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p\" (UID: \"43faa7ed-99a7-478c-8994-7a97333df231\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" Feb 25 13:35:35 crc kubenswrapper[4815]: I0225 13:35:35.522456 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43faa7ed-99a7-478c-8994-7a97333df231-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p\" (UID: \"43faa7ed-99a7-478c-8994-7a97333df231\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" Feb 25 13:35:35 crc kubenswrapper[4815]: I0225 13:35:35.559439 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" Feb 25 13:35:35 crc kubenswrapper[4815]: I0225 13:35:35.941395 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:35 crc kubenswrapper[4815]: I0225 13:35:35.941674 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:35 crc kubenswrapper[4815]: I0225 13:35:35.950025 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-metrics-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:35 crc kubenswrapper[4815]: I0225 13:35:35.950055 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/509894d6-7814-4f85-9eed-f1946b029dbf-webhook-certs\") pod \"openstack-operator-controller-manager-5f47dfdb6c-slw4h\" (UID: \"509894d6-7814-4f85-9eed-f1946b029dbf\") " pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:35 crc kubenswrapper[4815]: I0225 13:35:35.980919 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:36 crc kubenswrapper[4815]: I0225 13:35:36.030656 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p"] Feb 25 13:35:36 crc kubenswrapper[4815]: I0225 13:35:36.433315 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h"] Feb 25 13:35:36 crc kubenswrapper[4815]: I0225 13:35:36.578496 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" event={"ID":"509894d6-7814-4f85-9eed-f1946b029dbf","Type":"ContainerStarted","Data":"3972f057558b5f254e436980d14223508e17c0c8b6dfa99fb0175ac6c4027cc9"} Feb 25 13:35:36 crc kubenswrapper[4815]: I0225 13:35:36.580732 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" event={"ID":"43faa7ed-99a7-478c-8994-7a97333df231","Type":"ContainerStarted","Data":"a88b24849576e18b8c956b46c5878fe5dba2981934d6edfc13a337fe09375834"} Feb 25 13:35:37 crc kubenswrapper[4815]: I0225 13:35:37.595429 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" event={"ID":"509894d6-7814-4f85-9eed-f1946b029dbf","Type":"ContainerStarted","Data":"948b4454622215db407b483238757214a9bb1b38df1f85f5f506e8b8e5b15611"} Feb 25 13:35:37 crc kubenswrapper[4815]: I0225 13:35:37.595921 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:37 crc kubenswrapper[4815]: I0225 13:35:37.645650 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" podStartSLOduration=34.645617637 podStartE2EDuration="34.645617637s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:35:37.638451035 +0000 UTC m=+895.439549129" watchObservedRunningTime="2026-02-25 13:35:37.645617637 +0000 UTC m=+895.446715731" Feb 25 13:35:38 crc kubenswrapper[4815]: I0225 13:35:38.603957 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-8vr5p" event={"ID":"0673a594-2080-4325-80ce-f5597ad337c7","Type":"ContainerStarted","Data":"964dca74790d3dc6083fc0c905a0726b70ae684cc74cd99df07357a1b5af3059"} Feb 25 13:35:38 crc kubenswrapper[4815]: I0225 13:35:38.604362 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-68f46476f-8vr5p" Feb 25 13:35:38 crc kubenswrapper[4815]: I0225 13:35:38.605297 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-vnb57" event={"ID":"6bab4cd6-5eff-409c-83e3-ad5c1a6b6444","Type":"ContainerStarted","Data":"7ce1205311599027dfb6e69fc3c1e4040e962071392c384a220f44ef17ae5288"} Feb 25 13:35:38 crc kubenswrapper[4815]: I0225 13:35:38.605553 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-vnb57" Feb 25 13:35:38 crc kubenswrapper[4815]: I0225 13:35:38.606972 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-48hkj" event={"ID":"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4","Type":"ContainerStarted","Data":"777a253843af9b52b573948537ab50210ba54dbf69e9b38c140b76f89a3e3895"} Feb 25 13:35:38 crc kubenswrapper[4815]: I0225 13:35:38.627945 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-68f46476f-8vr5p" podStartSLOduration=2.551100721 podStartE2EDuration="35.627928287s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:05.20938474 +0000 UTC m=+863.010482794" lastFinishedPulling="2026-02-25 13:35:38.286212306 +0000 UTC m=+896.087310360" observedRunningTime="2026-02-25 13:35:38.625959617 +0000 UTC m=+896.427057681" watchObservedRunningTime="2026-02-25 13:35:38.627928287 +0000 UTC m=+896.429026341" Feb 25 13:35:39 crc kubenswrapper[4815]: I0225 13:35:39.354824 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-79d975b745-s5ztn" Feb 25 13:35:39 crc kubenswrapper[4815]: I0225 13:35:39.390813 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-vnb57" podStartSLOduration=3.295559891 podStartE2EDuration="36.390796856s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:05.190896009 +0000 UTC m=+862.991994063" lastFinishedPulling="2026-02-25 13:35:38.286132944 +0000 UTC m=+896.087231028" observedRunningTime="2026-02-25 13:35:38.665022064 +0000 UTC m=+896.466120128" watchObservedRunningTime="2026-02-25 13:35:39.390796856 +0000 UTC m=+897.191894900" Feb 25 13:35:39 crc kubenswrapper[4815]: I0225 13:35:39.625901 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-rwzjh" event={"ID":"136c32bc-0a85-4a36-b527-a787fa7158f1","Type":"ContainerStarted","Data":"b7c30d9c9c90685de9fd661cd2c6c6882511b85094ff0599b7141bbd2e20eaf8"} Feb 25 13:35:39 crc kubenswrapper[4815]: I0225 13:35:39.636227 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-rx4fg" event={"ID":"c64a07be-3cda-473e-89b6-fd0347519c93","Type":"ContainerStarted","Data":"374fdebd3db78824e50581f56e9de5f94c91ae84a1d196ddec3369e6ca4e41c8"} Feb 25 13:35:39 crc kubenswrapper[4815]: I0225 13:35:39.637356 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-rwzjh" Feb 25 13:35:39 crc kubenswrapper[4815]: I0225 13:35:39.637416 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-rx4fg" Feb 25 13:35:39 crc kubenswrapper[4815]: I0225 13:35:39.650252 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-zwknl" event={"ID":"62c00caf-05cf-4229-9349-341a2b597b12","Type":"ContainerStarted","Data":"9057569e9aa03328b7ab66e11edeef09b0cd55e4f96fdb27f29180f4be520ba1"} Feb 25 13:35:39 crc kubenswrapper[4815]: I0225 13:35:39.650712 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-zwknl" Feb 25 13:35:39 crc kubenswrapper[4815]: I0225 13:35:39.652381 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-z4jrr" event={"ID":"b30720e2-30f0-480e-a44f-2b409be766c6","Type":"ContainerStarted","Data":"ce5d0cd2b7e1a73711ca807e3801ffabc00f591ad647e743924d14e7a9f5a5dc"} Feb 25 13:35:39 crc kubenswrapper[4815]: I0225 13:35:39.653022 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-z4jrr" Feb 25 13:35:39 crc kubenswrapper[4815]: I0225 13:35:39.656893 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-rwzjh" podStartSLOduration=3.357224516 podStartE2EDuration="36.656881799s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:05.193496529 +0000 UTC m=+862.994594583" lastFinishedPulling="2026-02-25 13:35:38.493153812 +0000 UTC m=+896.294251866" observedRunningTime="2026-02-25 13:35:39.656454226 +0000 UTC m=+897.457552290" watchObservedRunningTime="2026-02-25 13:35:39.656881799 +0000 UTC m=+897.457979853" Feb 25 13:35:39 crc kubenswrapper[4815]: I0225 13:35:39.659096 4815 generic.go:334] "Generic (PLEG): container finished" podID="eec5ea92-0b0f-4c61-a0ee-074a5a21eba4" containerID="777a253843af9b52b573948537ab50210ba54dbf69e9b38c140b76f89a3e3895" exitCode=0 Feb 25 13:35:39 crc kubenswrapper[4815]: I0225 13:35:39.660362 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-48hkj" event={"ID":"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4","Type":"ContainerDied","Data":"777a253843af9b52b573948537ab50210ba54dbf69e9b38c140b76f89a3e3895"} Feb 25 13:35:39 crc kubenswrapper[4815]: I0225 13:35:39.676678 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-rx4fg" podStartSLOduration=3.376416218 podStartE2EDuration="36.67666131s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:05.191745795 +0000 UTC m=+862.992843849" lastFinishedPulling="2026-02-25 13:35:38.491990887 +0000 UTC m=+896.293088941" observedRunningTime="2026-02-25 13:35:39.673608167 +0000 UTC m=+897.474706221" watchObservedRunningTime="2026-02-25 13:35:39.67666131 +0000 UTC m=+897.477759364" Feb 25 13:35:39 crc kubenswrapper[4815]: I0225 13:35:39.690134 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-zwknl" podStartSLOduration=2.376195876 podStartE2EDuration="36.690116057s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:04.179291393 +0000 UTC m=+861.980389447" lastFinishedPulling="2026-02-25 13:35:38.493211574 +0000 UTC m=+896.294309628" observedRunningTime="2026-02-25 13:35:39.686863436 +0000 UTC m=+897.487961490" watchObservedRunningTime="2026-02-25 13:35:39.690116057 +0000 UTC m=+897.491214111" Feb 25 13:35:39 crc kubenswrapper[4815]: I0225 13:35:39.721679 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-z4jrr" podStartSLOduration=2.9977095350000003 podStartE2EDuration="36.721661821s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:04.767044179 +0000 UTC m=+862.568142233" lastFinishedPulling="2026-02-25 13:35:38.490996465 +0000 UTC m=+896.292094519" observedRunningTime="2026-02-25 13:35:39.71515049 +0000 UTC m=+897.516248544" watchObservedRunningTime="2026-02-25 13:35:39.721661821 +0000 UTC m=+897.522759875" Feb 25 13:35:40 crc kubenswrapper[4815]: I0225 13:35:40.670682 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" event={"ID":"43faa7ed-99a7-478c-8994-7a97333df231","Type":"ContainerStarted","Data":"ae3b3abc5d442bafcdc951a621397c5fe619f1534f8c9dcbf1f07a3e0e80c76b"} Feb 25 13:35:40 crc kubenswrapper[4815]: I0225 13:35:40.672158 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" Feb 25 13:35:40 crc kubenswrapper[4815]: I0225 13:35:40.702792 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" podStartSLOduration=33.494098687 podStartE2EDuration="37.702776675s" podCreationTimestamp="2026-02-25 13:35:03 +0000 UTC" firstStartedPulling="2026-02-25 13:35:36.043135709 +0000 UTC m=+893.844233763" lastFinishedPulling="2026-02-25 13:35:40.251813707 +0000 UTC m=+898.052911751" observedRunningTime="2026-02-25 13:35:40.699350458 +0000 UTC m=+898.500448512" watchObservedRunningTime="2026-02-25 13:35:40.702776675 +0000 UTC m=+898.503874729" Feb 25 13:35:41 crc kubenswrapper[4815]: I0225 13:35:41.679592 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-48hkj" event={"ID":"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4","Type":"ContainerStarted","Data":"f06a1d59bf0d79fac7ada1a583f162db7448af42d737bb43c22f409aad5f7584"} Feb 25 13:35:41 crc kubenswrapper[4815]: I0225 13:35:41.707119 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-48hkj" podStartSLOduration=18.582079531 podStartE2EDuration="28.707076674s" podCreationTimestamp="2026-02-25 13:35:13 +0000 UTC" firstStartedPulling="2026-02-25 13:35:30.437975931 +0000 UTC m=+888.239073985" lastFinishedPulling="2026-02-25 13:35:40.562973084 +0000 UTC m=+898.364071128" observedRunningTime="2026-02-25 13:35:41.705821286 +0000 UTC m=+899.506919340" watchObservedRunningTime="2026-02-25 13:35:41.707076674 +0000 UTC m=+899.508174748" Feb 25 13:35:43 crc kubenswrapper[4815]: I0225 13:35:43.502207 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-zwknl" Feb 25 13:35:43 crc kubenswrapper[4815]: I0225 13:35:43.643663 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-csfk2" Feb 25 13:35:43 crc kubenswrapper[4815]: I0225 13:35:43.706787 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-ccl5w" Feb 25 13:35:43 crc kubenswrapper[4815]: I0225 13:35:43.724893 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-48hkj" Feb 25 13:35:43 crc kubenswrapper[4815]: I0225 13:35:43.724946 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-48hkj" Feb 25 13:35:43 crc kubenswrapper[4815]: I0225 13:35:43.737885 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-z4jrr" Feb 25 13:35:43 crc kubenswrapper[4815]: I0225 13:35:43.779836 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-48hkj" Feb 25 13:35:43 crc kubenswrapper[4815]: I0225 13:35:43.786666 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-52psd" Feb 25 13:35:43 crc kubenswrapper[4815]: I0225 13:35:43.837277 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-62tw8" Feb 25 13:35:43 crc kubenswrapper[4815]: I0225 13:35:43.863440 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-rx4fg" Feb 25 13:35:43 crc kubenswrapper[4815]: I0225 13:35:43.866069 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-67d996989d-5fq96" Feb 25 13:35:43 crc kubenswrapper[4815]: I0225 13:35:43.883786 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-vnb57" Feb 25 13:35:44 crc kubenswrapper[4815]: I0225 13:35:44.046688 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-68f46476f-8vr5p" Feb 25 13:35:44 crc kubenswrapper[4815]: I0225 13:35:44.092627 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-rwzjh" Feb 25 13:35:45 crc kubenswrapper[4815]: I0225 13:35:45.565905 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p" Feb 25 13:35:45 crc kubenswrapper[4815]: I0225 13:35:45.994080 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5f47dfdb6c-slw4h" Feb 25 13:35:53 crc kubenswrapper[4815]: I0225 13:35:53.778804 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-48hkj" Feb 25 13:35:53 crc kubenswrapper[4815]: I0225 13:35:53.841419 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-48hkj"] Feb 25 13:35:53 crc kubenswrapper[4815]: I0225 13:35:53.842026 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-48hkj" podUID="eec5ea92-0b0f-4c61-a0ee-074a5a21eba4" containerName="registry-server" containerID="cri-o://f06a1d59bf0d79fac7ada1a583f162db7448af42d737bb43c22f409aad5f7584" gracePeriod=2 Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.236500 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-48hkj" Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.322430 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-utilities\") pod \"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4\" (UID: \"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4\") " Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.322505 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpxlb\" (UniqueName: \"kubernetes.io/projected/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-kube-api-access-gpxlb\") pod \"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4\" (UID: \"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4\") " Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.322768 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-catalog-content\") pod \"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4\" (UID: \"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4\") " Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.323336 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-utilities" (OuterVolumeSpecName: "utilities") pod "eec5ea92-0b0f-4c61-a0ee-074a5a21eba4" (UID: "eec5ea92-0b0f-4c61-a0ee-074a5a21eba4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.327120 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.328459 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-kube-api-access-gpxlb" (OuterVolumeSpecName: "kube-api-access-gpxlb") pod "eec5ea92-0b0f-4c61-a0ee-074a5a21eba4" (UID: "eec5ea92-0b0f-4c61-a0ee-074a5a21eba4"). InnerVolumeSpecName "kube-api-access-gpxlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.396677 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eec5ea92-0b0f-4c61-a0ee-074a5a21eba4" (UID: "eec5ea92-0b0f-4c61-a0ee-074a5a21eba4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.428758 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.428796 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpxlb\" (UniqueName: \"kubernetes.io/projected/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4-kube-api-access-gpxlb\") on node \"crc\" DevicePath \"\"" Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.825301 4815 generic.go:334] "Generic (PLEG): container finished" podID="eec5ea92-0b0f-4c61-a0ee-074a5a21eba4" containerID="f06a1d59bf0d79fac7ada1a583f162db7448af42d737bb43c22f409aad5f7584" exitCode=0 Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.825373 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-48hkj" Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.825391 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-48hkj" event={"ID":"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4","Type":"ContainerDied","Data":"f06a1d59bf0d79fac7ada1a583f162db7448af42d737bb43c22f409aad5f7584"} Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.826798 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-48hkj" event={"ID":"eec5ea92-0b0f-4c61-a0ee-074a5a21eba4","Type":"ContainerDied","Data":"3b8d8d90fe0de3ebbd4ca814ca65569cc347b957e67a804e5cc0046ffe228d50"} Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.826837 4815 scope.go:117] "RemoveContainer" containerID="f06a1d59bf0d79fac7ada1a583f162db7448af42d737bb43c22f409aad5f7584" Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.853211 4815 scope.go:117] "RemoveContainer" containerID="777a253843af9b52b573948537ab50210ba54dbf69e9b38c140b76f89a3e3895" Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.881040 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-48hkj"] Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.890771 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-48hkj"] Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.905638 4815 scope.go:117] "RemoveContainer" containerID="863f30e473201d0c095c68514d1f388d7d6877e98546f2ed75acb86dbf6b17af" Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.925858 4815 scope.go:117] "RemoveContainer" containerID="f06a1d59bf0d79fac7ada1a583f162db7448af42d737bb43c22f409aad5f7584" Feb 25 13:35:54 crc kubenswrapper[4815]: E0225 13:35:54.926234 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f06a1d59bf0d79fac7ada1a583f162db7448af42d737bb43c22f409aad5f7584\": container with ID starting with f06a1d59bf0d79fac7ada1a583f162db7448af42d737bb43c22f409aad5f7584 not found: ID does not exist" containerID="f06a1d59bf0d79fac7ada1a583f162db7448af42d737bb43c22f409aad5f7584" Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.926335 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f06a1d59bf0d79fac7ada1a583f162db7448af42d737bb43c22f409aad5f7584"} err="failed to get container status \"f06a1d59bf0d79fac7ada1a583f162db7448af42d737bb43c22f409aad5f7584\": rpc error: code = NotFound desc = could not find container \"f06a1d59bf0d79fac7ada1a583f162db7448af42d737bb43c22f409aad5f7584\": container with ID starting with f06a1d59bf0d79fac7ada1a583f162db7448af42d737bb43c22f409aad5f7584 not found: ID does not exist" Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.926372 4815 scope.go:117] "RemoveContainer" containerID="777a253843af9b52b573948537ab50210ba54dbf69e9b38c140b76f89a3e3895" Feb 25 13:35:54 crc kubenswrapper[4815]: E0225 13:35:54.926864 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"777a253843af9b52b573948537ab50210ba54dbf69e9b38c140b76f89a3e3895\": container with ID starting with 777a253843af9b52b573948537ab50210ba54dbf69e9b38c140b76f89a3e3895 not found: ID does not exist" containerID="777a253843af9b52b573948537ab50210ba54dbf69e9b38c140b76f89a3e3895" Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.926900 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"777a253843af9b52b573948537ab50210ba54dbf69e9b38c140b76f89a3e3895"} err="failed to get container status \"777a253843af9b52b573948537ab50210ba54dbf69e9b38c140b76f89a3e3895\": rpc error: code = NotFound desc = could not find container \"777a253843af9b52b573948537ab50210ba54dbf69e9b38c140b76f89a3e3895\": container with ID starting with 777a253843af9b52b573948537ab50210ba54dbf69e9b38c140b76f89a3e3895 not found: ID does not exist" Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.926921 4815 scope.go:117] "RemoveContainer" containerID="863f30e473201d0c095c68514d1f388d7d6877e98546f2ed75acb86dbf6b17af" Feb 25 13:35:54 crc kubenswrapper[4815]: E0225 13:35:54.927361 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"863f30e473201d0c095c68514d1f388d7d6877e98546f2ed75acb86dbf6b17af\": container with ID starting with 863f30e473201d0c095c68514d1f388d7d6877e98546f2ed75acb86dbf6b17af not found: ID does not exist" containerID="863f30e473201d0c095c68514d1f388d7d6877e98546f2ed75acb86dbf6b17af" Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.927407 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"863f30e473201d0c095c68514d1f388d7d6877e98546f2ed75acb86dbf6b17af"} err="failed to get container status \"863f30e473201d0c095c68514d1f388d7d6877e98546f2ed75acb86dbf6b17af\": rpc error: code = NotFound desc = could not find container \"863f30e473201d0c095c68514d1f388d7d6877e98546f2ed75acb86dbf6b17af\": container with ID starting with 863f30e473201d0c095c68514d1f388d7d6877e98546f2ed75acb86dbf6b17af not found: ID does not exist" Feb 25 13:35:54 crc kubenswrapper[4815]: I0225 13:35:54.947668 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eec5ea92-0b0f-4c61-a0ee-074a5a21eba4" path="/var/lib/kubelet/pods/eec5ea92-0b0f-4c61-a0ee-074a5a21eba4/volumes" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.155296 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533776-kz9qn"] Feb 25 13:36:00 crc kubenswrapper[4815]: E0225 13:36:00.156566 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff19dbc-fec6-46e8-b36a-e49dde7db51c" containerName="registry-server" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.156600 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff19dbc-fec6-46e8-b36a-e49dde7db51c" containerName="registry-server" Feb 25 13:36:00 crc kubenswrapper[4815]: E0225 13:36:00.156627 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59dc6b8b-e3e2-4f25-91bf-310d17bbd603" containerName="extract-utilities" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.156646 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="59dc6b8b-e3e2-4f25-91bf-310d17bbd603" containerName="extract-utilities" Feb 25 13:36:00 crc kubenswrapper[4815]: E0225 13:36:00.156705 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59dc6b8b-e3e2-4f25-91bf-310d17bbd603" containerName="extract-content" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.156723 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="59dc6b8b-e3e2-4f25-91bf-310d17bbd603" containerName="extract-content" Feb 25 13:36:00 crc kubenswrapper[4815]: E0225 13:36:00.156747 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff19dbc-fec6-46e8-b36a-e49dde7db51c" containerName="extract-content" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.156763 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff19dbc-fec6-46e8-b36a-e49dde7db51c" containerName="extract-content" Feb 25 13:36:00 crc kubenswrapper[4815]: E0225 13:36:00.156794 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec5ea92-0b0f-4c61-a0ee-074a5a21eba4" containerName="registry-server" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.156809 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec5ea92-0b0f-4c61-a0ee-074a5a21eba4" containerName="registry-server" Feb 25 13:36:00 crc kubenswrapper[4815]: E0225 13:36:00.156839 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59dc6b8b-e3e2-4f25-91bf-310d17bbd603" containerName="registry-server" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.156854 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="59dc6b8b-e3e2-4f25-91bf-310d17bbd603" containerName="registry-server" Feb 25 13:36:00 crc kubenswrapper[4815]: E0225 13:36:00.156873 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec5ea92-0b0f-4c61-a0ee-074a5a21eba4" containerName="extract-utilities" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.156889 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec5ea92-0b0f-4c61-a0ee-074a5a21eba4" containerName="extract-utilities" Feb 25 13:36:00 crc kubenswrapper[4815]: E0225 13:36:00.156916 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec5ea92-0b0f-4c61-a0ee-074a5a21eba4" containerName="extract-content" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.156932 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec5ea92-0b0f-4c61-a0ee-074a5a21eba4" containerName="extract-content" Feb 25 13:36:00 crc kubenswrapper[4815]: E0225 13:36:00.156971 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff19dbc-fec6-46e8-b36a-e49dde7db51c" containerName="extract-utilities" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.156987 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff19dbc-fec6-46e8-b36a-e49dde7db51c" containerName="extract-utilities" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.157296 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="eec5ea92-0b0f-4c61-a0ee-074a5a21eba4" containerName="registry-server" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.157341 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="59dc6b8b-e3e2-4f25-91bf-310d17bbd603" containerName="registry-server" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.157357 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff19dbc-fec6-46e8-b36a-e49dde7db51c" containerName="registry-server" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.158234 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533776-kz9qn" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.161120 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.161319 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.164621 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.164899 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533776-kz9qn"] Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.213215 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5vgg\" (UniqueName: \"kubernetes.io/projected/805e511a-ead5-45ec-bcdc-018a16e7f059-kube-api-access-z5vgg\") pod \"auto-csr-approver-29533776-kz9qn\" (UID: \"805e511a-ead5-45ec-bcdc-018a16e7f059\") " pod="openshift-infra/auto-csr-approver-29533776-kz9qn" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.314899 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5vgg\" (UniqueName: \"kubernetes.io/projected/805e511a-ead5-45ec-bcdc-018a16e7f059-kube-api-access-z5vgg\") pod \"auto-csr-approver-29533776-kz9qn\" (UID: \"805e511a-ead5-45ec-bcdc-018a16e7f059\") " pod="openshift-infra/auto-csr-approver-29533776-kz9qn" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.348169 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5vgg\" (UniqueName: \"kubernetes.io/projected/805e511a-ead5-45ec-bcdc-018a16e7f059-kube-api-access-z5vgg\") pod \"auto-csr-approver-29533776-kz9qn\" (UID: \"805e511a-ead5-45ec-bcdc-018a16e7f059\") " pod="openshift-infra/auto-csr-approver-29533776-kz9qn" Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.487675 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533776-kz9qn" Feb 25 13:36:00 crc kubenswrapper[4815]: W0225 13:36:00.947817 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod805e511a_ead5_45ec_bcdc_018a16e7f059.slice/crio-1521d7c4bb221c63e55a5697f95c4e2f6acded6e729a6c2c3125114b9d617e10 WatchSource:0}: Error finding container 1521d7c4bb221c63e55a5697f95c4e2f6acded6e729a6c2c3125114b9d617e10: Status 404 returned error can't find the container with id 1521d7c4bb221c63e55a5697f95c4e2f6acded6e729a6c2c3125114b9d617e10 Feb 25 13:36:00 crc kubenswrapper[4815]: I0225 13:36:00.948285 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533776-kz9qn"] Feb 25 13:36:01 crc kubenswrapper[4815]: I0225 13:36:01.888403 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533776-kz9qn" event={"ID":"805e511a-ead5-45ec-bcdc-018a16e7f059","Type":"ContainerStarted","Data":"1521d7c4bb221c63e55a5697f95c4e2f6acded6e729a6c2c3125114b9d617e10"} Feb 25 13:36:02 crc kubenswrapper[4815]: I0225 13:36:02.841856 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-74q86"] Feb 25 13:36:02 crc kubenswrapper[4815]: I0225 13:36:02.842935 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-74q86" Feb 25 13:36:02 crc kubenswrapper[4815]: I0225 13:36:02.847494 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 25 13:36:02 crc kubenswrapper[4815]: I0225 13:36:02.847753 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 25 13:36:02 crc kubenswrapper[4815]: I0225 13:36:02.847909 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-ms8rp" Feb 25 13:36:02 crc kubenswrapper[4815]: I0225 13:36:02.848057 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 25 13:36:02 crc kubenswrapper[4815]: I0225 13:36:02.857952 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-74q86"] Feb 25 13:36:02 crc kubenswrapper[4815]: I0225 13:36:02.914407 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-npxbr"] Feb 25 13:36:02 crc kubenswrapper[4815]: I0225 13:36:02.915815 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-npxbr" Feb 25 13:36:02 crc kubenswrapper[4815]: I0225 13:36:02.917840 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 25 13:36:02 crc kubenswrapper[4815]: I0225 13:36:02.951243 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86dqs\" (UniqueName: \"kubernetes.io/projected/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-kube-api-access-86dqs\") pod \"dnsmasq-dns-78dd6ddcc-npxbr\" (UID: \"81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-npxbr" Feb 25 13:36:02 crc kubenswrapper[4815]: I0225 13:36:02.951308 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-config\") pod \"dnsmasq-dns-78dd6ddcc-npxbr\" (UID: \"81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-npxbr" Feb 25 13:36:02 crc kubenswrapper[4815]: I0225 13:36:02.951333 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-npxbr\" (UID: \"81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-npxbr" Feb 25 13:36:02 crc kubenswrapper[4815]: I0225 13:36:02.951361 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11720769-5e77-4722-bfba-f208fcc3276d-config\") pod \"dnsmasq-dns-675f4bcbfc-74q86\" (UID: \"11720769-5e77-4722-bfba-f208fcc3276d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-74q86" Feb 25 13:36:02 crc kubenswrapper[4815]: I0225 13:36:02.951455 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh528\" (UniqueName: \"kubernetes.io/projected/11720769-5e77-4722-bfba-f208fcc3276d-kube-api-access-mh528\") pod \"dnsmasq-dns-675f4bcbfc-74q86\" (UID: \"11720769-5e77-4722-bfba-f208fcc3276d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-74q86" Feb 25 13:36:02 crc kubenswrapper[4815]: I0225 13:36:02.974190 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-npxbr"] Feb 25 13:36:03 crc kubenswrapper[4815]: I0225 13:36:03.053026 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86dqs\" (UniqueName: \"kubernetes.io/projected/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-kube-api-access-86dqs\") pod \"dnsmasq-dns-78dd6ddcc-npxbr\" (UID: \"81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-npxbr" Feb 25 13:36:03 crc kubenswrapper[4815]: I0225 13:36:03.053298 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-npxbr\" (UID: \"81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-npxbr" Feb 25 13:36:03 crc kubenswrapper[4815]: I0225 13:36:03.053653 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-config\") pod \"dnsmasq-dns-78dd6ddcc-npxbr\" (UID: \"81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-npxbr" Feb 25 13:36:03 crc kubenswrapper[4815]: I0225 13:36:03.054026 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11720769-5e77-4722-bfba-f208fcc3276d-config\") pod \"dnsmasq-dns-675f4bcbfc-74q86\" (UID: \"11720769-5e77-4722-bfba-f208fcc3276d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-74q86" Feb 25 13:36:03 crc kubenswrapper[4815]: I0225 13:36:03.054210 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh528\" (UniqueName: \"kubernetes.io/projected/11720769-5e77-4722-bfba-f208fcc3276d-kube-api-access-mh528\") pod \"dnsmasq-dns-675f4bcbfc-74q86\" (UID: \"11720769-5e77-4722-bfba-f208fcc3276d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-74q86" Feb 25 13:36:03 crc kubenswrapper[4815]: I0225 13:36:03.054453 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-npxbr\" (UID: \"81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-npxbr" Feb 25 13:36:03 crc kubenswrapper[4815]: I0225 13:36:03.054814 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-config\") pod \"dnsmasq-dns-78dd6ddcc-npxbr\" (UID: \"81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-npxbr" Feb 25 13:36:03 crc kubenswrapper[4815]: I0225 13:36:03.054820 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11720769-5e77-4722-bfba-f208fcc3276d-config\") pod \"dnsmasq-dns-675f4bcbfc-74q86\" (UID: \"11720769-5e77-4722-bfba-f208fcc3276d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-74q86" Feb 25 13:36:03 crc kubenswrapper[4815]: I0225 13:36:03.072679 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86dqs\" (UniqueName: \"kubernetes.io/projected/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-kube-api-access-86dqs\") pod \"dnsmasq-dns-78dd6ddcc-npxbr\" (UID: \"81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-npxbr" Feb 25 13:36:03 crc kubenswrapper[4815]: I0225 13:36:03.072872 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh528\" (UniqueName: \"kubernetes.io/projected/11720769-5e77-4722-bfba-f208fcc3276d-kube-api-access-mh528\") pod \"dnsmasq-dns-675f4bcbfc-74q86\" (UID: \"11720769-5e77-4722-bfba-f208fcc3276d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-74q86" Feb 25 13:36:03 crc kubenswrapper[4815]: I0225 13:36:03.161119 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-74q86" Feb 25 13:36:03 crc kubenswrapper[4815]: I0225 13:36:03.232080 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-npxbr" Feb 25 13:36:04 crc kubenswrapper[4815]: I0225 13:36:04.190593 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-74q86"] Feb 25 13:36:04 crc kubenswrapper[4815]: W0225 13:36:04.200746 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11720769_5e77_4722_bfba_f208fcc3276d.slice/crio-8b27f1495338a9f8d2297103bbcae67f651a485118e4544bce4a7025aafcdec8 WatchSource:0}: Error finding container 8b27f1495338a9f8d2297103bbcae67f651a485118e4544bce4a7025aafcdec8: Status 404 returned error can't find the container with id 8b27f1495338a9f8d2297103bbcae67f651a485118e4544bce4a7025aafcdec8 Feb 25 13:36:04 crc kubenswrapper[4815]: I0225 13:36:04.244744 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-npxbr"] Feb 25 13:36:04 crc kubenswrapper[4815]: I0225 13:36:04.914786 4815 generic.go:334] "Generic (PLEG): container finished" podID="805e511a-ead5-45ec-bcdc-018a16e7f059" containerID="555b0db9fc48103d6923e1891f56a10d3ab01f58a617ce1367f71c3820cfa3b0" exitCode=0 Feb 25 13:36:04 crc kubenswrapper[4815]: I0225 13:36:04.915227 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533776-kz9qn" event={"ID":"805e511a-ead5-45ec-bcdc-018a16e7f059","Type":"ContainerDied","Data":"555b0db9fc48103d6923e1891f56a10d3ab01f58a617ce1367f71c3820cfa3b0"} Feb 25 13:36:04 crc kubenswrapper[4815]: I0225 13:36:04.918741 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-npxbr" event={"ID":"81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6","Type":"ContainerStarted","Data":"e64591df5903b5ddf98ae24578aea53c1f5957213ad4fec14b34ad5d24e05c10"} Feb 25 13:36:04 crc kubenswrapper[4815]: I0225 13:36:04.920260 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-74q86" event={"ID":"11720769-5e77-4722-bfba-f208fcc3276d","Type":"ContainerStarted","Data":"8b27f1495338a9f8d2297103bbcae67f651a485118e4544bce4a7025aafcdec8"} Feb 25 13:36:05 crc kubenswrapper[4815]: I0225 13:36:05.719187 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-74q86"] Feb 25 13:36:05 crc kubenswrapper[4815]: I0225 13:36:05.754098 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ktjsd"] Feb 25 13:36:05 crc kubenswrapper[4815]: I0225 13:36:05.755207 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-ktjsd" Feb 25 13:36:05 crc kubenswrapper[4815]: I0225 13:36:05.774152 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ktjsd"] Feb 25 13:36:05 crc kubenswrapper[4815]: I0225 13:36:05.814211 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwtwf\" (UniqueName: \"kubernetes.io/projected/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-kube-api-access-lwtwf\") pod \"dnsmasq-dns-666b6646f7-ktjsd\" (UID: \"9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5\") " pod="openstack/dnsmasq-dns-666b6646f7-ktjsd" Feb 25 13:36:05 crc kubenswrapper[4815]: I0225 13:36:05.814311 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-config\") pod \"dnsmasq-dns-666b6646f7-ktjsd\" (UID: \"9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5\") " pod="openstack/dnsmasq-dns-666b6646f7-ktjsd" Feb 25 13:36:05 crc kubenswrapper[4815]: I0225 13:36:05.814338 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-ktjsd\" (UID: \"9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5\") " pod="openstack/dnsmasq-dns-666b6646f7-ktjsd" Feb 25 13:36:05 crc kubenswrapper[4815]: I0225 13:36:05.917363 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwtwf\" (UniqueName: \"kubernetes.io/projected/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-kube-api-access-lwtwf\") pod \"dnsmasq-dns-666b6646f7-ktjsd\" (UID: \"9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5\") " pod="openstack/dnsmasq-dns-666b6646f7-ktjsd" Feb 25 13:36:05 crc kubenswrapper[4815]: I0225 13:36:05.917490 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-config\") pod \"dnsmasq-dns-666b6646f7-ktjsd\" (UID: \"9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5\") " pod="openstack/dnsmasq-dns-666b6646f7-ktjsd" Feb 25 13:36:05 crc kubenswrapper[4815]: I0225 13:36:05.917524 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-ktjsd\" (UID: \"9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5\") " pod="openstack/dnsmasq-dns-666b6646f7-ktjsd" Feb 25 13:36:05 crc kubenswrapper[4815]: I0225 13:36:05.918372 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-config\") pod \"dnsmasq-dns-666b6646f7-ktjsd\" (UID: \"9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5\") " pod="openstack/dnsmasq-dns-666b6646f7-ktjsd" Feb 25 13:36:05 crc kubenswrapper[4815]: I0225 13:36:05.918372 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-dns-svc\") pod \"dnsmasq-dns-666b6646f7-ktjsd\" (UID: \"9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5\") " pod="openstack/dnsmasq-dns-666b6646f7-ktjsd" Feb 25 13:36:05 crc kubenswrapper[4815]: I0225 13:36:05.962006 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwtwf\" (UniqueName: \"kubernetes.io/projected/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-kube-api-access-lwtwf\") pod \"dnsmasq-dns-666b6646f7-ktjsd\" (UID: \"9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5\") " pod="openstack/dnsmasq-dns-666b6646f7-ktjsd" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.048866 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-npxbr"] Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.073291 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-ktjsd" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.096627 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmkxb"] Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.097782 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bmkxb" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.116084 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmkxb"] Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.222796 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb1d381-cd64-457f-b81a-49851aae7c68-config\") pod \"dnsmasq-dns-57d769cc4f-bmkxb\" (UID: \"9bb1d381-cd64-457f-b81a-49851aae7c68\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkxb" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.222855 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp8ns\" (UniqueName: \"kubernetes.io/projected/9bb1d381-cd64-457f-b81a-49851aae7c68-kube-api-access-tp8ns\") pod \"dnsmasq-dns-57d769cc4f-bmkxb\" (UID: \"9bb1d381-cd64-457f-b81a-49851aae7c68\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkxb" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.222893 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bb1d381-cd64-457f-b81a-49851aae7c68-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bmkxb\" (UID: \"9bb1d381-cd64-457f-b81a-49851aae7c68\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkxb" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.277736 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533776-kz9qn" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.324203 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp8ns\" (UniqueName: \"kubernetes.io/projected/9bb1d381-cd64-457f-b81a-49851aae7c68-kube-api-access-tp8ns\") pod \"dnsmasq-dns-57d769cc4f-bmkxb\" (UID: \"9bb1d381-cd64-457f-b81a-49851aae7c68\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkxb" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.324258 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bb1d381-cd64-457f-b81a-49851aae7c68-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bmkxb\" (UID: \"9bb1d381-cd64-457f-b81a-49851aae7c68\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkxb" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.324363 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb1d381-cd64-457f-b81a-49851aae7c68-config\") pod \"dnsmasq-dns-57d769cc4f-bmkxb\" (UID: \"9bb1d381-cd64-457f-b81a-49851aae7c68\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkxb" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.325117 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bb1d381-cd64-457f-b81a-49851aae7c68-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bmkxb\" (UID: \"9bb1d381-cd64-457f-b81a-49851aae7c68\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkxb" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.328611 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb1d381-cd64-457f-b81a-49851aae7c68-config\") pod \"dnsmasq-dns-57d769cc4f-bmkxb\" (UID: \"9bb1d381-cd64-457f-b81a-49851aae7c68\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkxb" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.353461 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp8ns\" (UniqueName: \"kubernetes.io/projected/9bb1d381-cd64-457f-b81a-49851aae7c68-kube-api-access-tp8ns\") pod \"dnsmasq-dns-57d769cc4f-bmkxb\" (UID: \"9bb1d381-cd64-457f-b81a-49851aae7c68\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkxb" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.423477 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ktjsd"] Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.426360 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5vgg\" (UniqueName: \"kubernetes.io/projected/805e511a-ead5-45ec-bcdc-018a16e7f059-kube-api-access-z5vgg\") pod \"805e511a-ead5-45ec-bcdc-018a16e7f059\" (UID: \"805e511a-ead5-45ec-bcdc-018a16e7f059\") " Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.429779 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/805e511a-ead5-45ec-bcdc-018a16e7f059-kube-api-access-z5vgg" (OuterVolumeSpecName: "kube-api-access-z5vgg") pod "805e511a-ead5-45ec-bcdc-018a16e7f059" (UID: "805e511a-ead5-45ec-bcdc-018a16e7f059"). InnerVolumeSpecName "kube-api-access-z5vgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.473726 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bmkxb" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.527556 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5vgg\" (UniqueName: \"kubernetes.io/projected/805e511a-ead5-45ec-bcdc-018a16e7f059-kube-api-access-z5vgg\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.670039 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmkxb"] Feb 25 13:36:06 crc kubenswrapper[4815]: W0225 13:36:06.678293 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9bb1d381_cd64_457f_b81a_49851aae7c68.slice/crio-4835449f02f41a94f4c8339cfe1cc5fbe8c637d8313506c73a373e385e49219b WatchSource:0}: Error finding container 4835449f02f41a94f4c8339cfe1cc5fbe8c637d8313506c73a373e385e49219b: Status 404 returned error can't find the container with id 4835449f02f41a94f4c8339cfe1cc5fbe8c637d8313506c73a373e385e49219b Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.910309 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 13:36:06 crc kubenswrapper[4815]: E0225 13:36:06.910639 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="805e511a-ead5-45ec-bcdc-018a16e7f059" containerName="oc" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.910653 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="805e511a-ead5-45ec-bcdc-018a16e7f059" containerName="oc" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.910796 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="805e511a-ead5-45ec-bcdc-018a16e7f059" containerName="oc" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.911454 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.913943 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.914062 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.914094 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.914409 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-2m84j" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.914618 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.914797 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.919247 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.926139 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 13:36:06 crc kubenswrapper[4815]: I0225 13:36:06.977168 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533776-kz9qn" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:06.999635 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bmkxb" event={"ID":"9bb1d381-cd64-457f-b81a-49851aae7c68","Type":"ContainerStarted","Data":"4835449f02f41a94f4c8339cfe1cc5fbe8c637d8313506c73a373e385e49219b"} Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:06.999681 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-ktjsd" event={"ID":"9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5","Type":"ContainerStarted","Data":"c424a1af5ed8fdede4017fd380fef7224338f87c27113a3c7b11ddfe9f2864cc"} Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:06.999698 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533776-kz9qn" event={"ID":"805e511a-ead5-45ec-bcdc-018a16e7f059","Type":"ContainerDied","Data":"1521d7c4bb221c63e55a5697f95c4e2f6acded6e729a6c2c3125114b9d617e10"} Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:06.999714 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1521d7c4bb221c63e55a5697f95c4e2f6acded6e729a6c2c3125114b9d617e10" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.065634 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsfhb\" (UniqueName: \"kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-kube-api-access-zsfhb\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.065702 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.065767 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.065790 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.065839 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.065857 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.065884 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.065926 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.065951 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.065991 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-config-data\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.066018 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.166895 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.166971 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsfhb\" (UniqueName: \"kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-kube-api-access-zsfhb\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.167006 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.167043 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.167065 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.167103 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.167124 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.167152 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.167183 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.167210 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.167237 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-config-data\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.168130 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-config-data\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.168194 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.169087 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.173230 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.173864 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.174141 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.174351 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: E0225 13:36:07.230199 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod805e511a_ead5_45ec_bcdc_018a16e7f059.slice\": RecentStats: unable to find data in memory cache]" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.230234 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.230730 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.233146 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.243222 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsfhb\" (UniqueName: \"kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-kube-api-access-zsfhb\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.245802 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.248262 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.249408 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.253284 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.253602 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.255224 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.255421 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.255543 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-js72s" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.255655 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.259954 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.264539 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.329863 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.348065 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533770-qrbhh"] Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.352679 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533770-qrbhh"] Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.370340 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.370411 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.370434 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0b98e7b2-3068-4f59-b25f-4755bd752b49-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.370453 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0b98e7b2-3068-4f59-b25f-4755bd752b49-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.370471 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.370485 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.370520 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.370539 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.370557 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.370583 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.370598 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9gpj\" (UniqueName: \"kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-kube-api-access-h9gpj\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.471500 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.471592 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.471613 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0b98e7b2-3068-4f59-b25f-4755bd752b49-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.471635 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0b98e7b2-3068-4f59-b25f-4755bd752b49-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.471652 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.471672 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.471690 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.471708 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.471746 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.471774 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.471789 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9gpj\" (UniqueName: \"kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-kube-api-access-h9gpj\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.472412 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.473287 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.473322 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.474082 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.474333 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.479001 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0b98e7b2-3068-4f59-b25f-4755bd752b49-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.480488 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.508484 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0b98e7b2-3068-4f59-b25f-4755bd752b49-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.508567 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.513462 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.517932 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9gpj\" (UniqueName: \"kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-kube-api-access-h9gpj\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.518549 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.603292 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.699898 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 13:36:07 crc kubenswrapper[4815]: W0225 13:36:07.702580 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a91fa1d_2609_4b30_9ea1_8a5b10145c8d.slice/crio-7949d637d954a9850bc8c2a36f253eb23f291f7be5db6df97664f033f9b4163f WatchSource:0}: Error finding container 7949d637d954a9850bc8c2a36f253eb23f291f7be5db6df97664f033f9b4163f: Status 404 returned error can't find the container with id 7949d637d954a9850bc8c2a36f253eb23f291f7be5db6df97664f033f9b4163f Feb 25 13:36:07 crc kubenswrapper[4815]: I0225 13:36:07.990078 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d","Type":"ContainerStarted","Data":"7949d637d954a9850bc8c2a36f253eb23f291f7be5db6df97664f033f9b4163f"} Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.091606 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 13:36:08 crc kubenswrapper[4815]: W0225 13:36:08.100786 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b98e7b2_3068_4f59_b25f_4755bd752b49.slice/crio-69a9e634865d2724b24095ef53b1d262fd17ab671b5c78bd4ad7379533e7b58c WatchSource:0}: Error finding container 69a9e634865d2724b24095ef53b1d262fd17ab671b5c78bd4ad7379533e7b58c: Status 404 returned error can't find the container with id 69a9e634865d2724b24095ef53b1d262fd17ab671b5c78bd4ad7379533e7b58c Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.225190 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.226481 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.230101 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.230236 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.237225 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.237284 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-h6ljw" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.237923 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.241193 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.285094 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-config-data-generated\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.285137 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v4zz\" (UniqueName: \"kubernetes.io/projected/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-kube-api-access-7v4zz\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.285174 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.285195 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.285213 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-kolla-config\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.285234 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.285254 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-operator-scripts\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.285278 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-config-data-default\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.386876 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-config-data-generated\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.386931 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v4zz\" (UniqueName: \"kubernetes.io/projected/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-kube-api-access-7v4zz\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.386980 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.387006 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.387027 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-kolla-config\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.387062 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.387091 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-operator-scripts\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.387124 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-config-data-default\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.387405 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.387456 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-config-data-generated\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.387787 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-kolla-config\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.388177 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-config-data-default\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.389404 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-operator-scripts\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.392933 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.404012 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v4zz\" (UniqueName: \"kubernetes.io/projected/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-kube-api-access-7v4zz\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.406181 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52d1a204-9bb6-40e1-a63c-09ffe6016cd9-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.409672 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"52d1a204-9bb6-40e1-a63c-09ffe6016cd9\") " pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.560037 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 25 13:36:08 crc kubenswrapper[4815]: I0225 13:36:08.945166 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67c273aa-5ff2-43fc-871f-64caafce5463" path="/var/lib/kubelet/pods/67c273aa-5ff2-43fc-871f-64caafce5463/volumes" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.000486 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0b98e7b2-3068-4f59-b25f-4755bd752b49","Type":"ContainerStarted","Data":"69a9e634865d2724b24095ef53b1d262fd17ab671b5c78bd4ad7379533e7b58c"} Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.749410 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.750566 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.754168 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-l8f4n" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.754547 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.754754 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.754909 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.760274 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.911366 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.911416 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.911458 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwj8l\" (UniqueName: \"kubernetes.io/projected/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-kube-api-access-qwj8l\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.911479 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.911500 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.911597 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.911623 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.911648 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.991819 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.992721 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.996645 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.996927 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.997107 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-k6nt2" Feb 25 13:36:09 crc kubenswrapper[4815]: I0225 13:36:09.997381 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.027054 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwj8l\" (UniqueName: \"kubernetes.io/projected/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-kube-api-access-qwj8l\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.027092 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.027113 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.027136 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/90528e4b-0537-4477-b7d1-603c21bc9358-kolla-config\") pod \"memcached-0\" (UID: \"90528e4b-0537-4477-b7d1-603c21bc9358\") " pod="openstack/memcached-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.027176 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.027195 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.027212 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.027274 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/90528e4b-0537-4477-b7d1-603c21bc9358-memcached-tls-certs\") pod \"memcached-0\" (UID: \"90528e4b-0537-4477-b7d1-603c21bc9358\") " pod="openstack/memcached-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.027311 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90528e4b-0537-4477-b7d1-603c21bc9358-combined-ca-bundle\") pod \"memcached-0\" (UID: \"90528e4b-0537-4477-b7d1-603c21bc9358\") " pod="openstack/memcached-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.027335 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpn7t\" (UniqueName: \"kubernetes.io/projected/90528e4b-0537-4477-b7d1-603c21bc9358-kube-api-access-rpn7t\") pod \"memcached-0\" (UID: \"90528e4b-0537-4477-b7d1-603c21bc9358\") " pod="openstack/memcached-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.027354 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90528e4b-0537-4477-b7d1-603c21bc9358-config-data\") pod \"memcached-0\" (UID: \"90528e4b-0537-4477-b7d1-603c21bc9358\") " pod="openstack/memcached-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.027375 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.027393 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.027788 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.028232 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.028343 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.028983 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.038126 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.038605 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.039876 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.044572 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwj8l\" (UniqueName: \"kubernetes.io/projected/c96edb5d-f9a9-4dc1-a9c2-e3def3efb850-kube-api-access-qwj8l\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.050261 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850\") " pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.072387 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.129752 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/90528e4b-0537-4477-b7d1-603c21bc9358-memcached-tls-certs\") pod \"memcached-0\" (UID: \"90528e4b-0537-4477-b7d1-603c21bc9358\") " pod="openstack/memcached-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.130030 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90528e4b-0537-4477-b7d1-603c21bc9358-combined-ca-bundle\") pod \"memcached-0\" (UID: \"90528e4b-0537-4477-b7d1-603c21bc9358\") " pod="openstack/memcached-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.130062 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpn7t\" (UniqueName: \"kubernetes.io/projected/90528e4b-0537-4477-b7d1-603c21bc9358-kube-api-access-rpn7t\") pod \"memcached-0\" (UID: \"90528e4b-0537-4477-b7d1-603c21bc9358\") " pod="openstack/memcached-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.130084 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90528e4b-0537-4477-b7d1-603c21bc9358-config-data\") pod \"memcached-0\" (UID: \"90528e4b-0537-4477-b7d1-603c21bc9358\") " pod="openstack/memcached-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.130139 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/90528e4b-0537-4477-b7d1-603c21bc9358-kolla-config\") pod \"memcached-0\" (UID: \"90528e4b-0537-4477-b7d1-603c21bc9358\") " pod="openstack/memcached-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.130935 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/90528e4b-0537-4477-b7d1-603c21bc9358-config-data\") pod \"memcached-0\" (UID: \"90528e4b-0537-4477-b7d1-603c21bc9358\") " pod="openstack/memcached-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.130972 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/90528e4b-0537-4477-b7d1-603c21bc9358-kolla-config\") pod \"memcached-0\" (UID: \"90528e4b-0537-4477-b7d1-603c21bc9358\") " pod="openstack/memcached-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.133353 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90528e4b-0537-4477-b7d1-603c21bc9358-combined-ca-bundle\") pod \"memcached-0\" (UID: \"90528e4b-0537-4477-b7d1-603c21bc9358\") " pod="openstack/memcached-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.146137 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/90528e4b-0537-4477-b7d1-603c21bc9358-memcached-tls-certs\") pod \"memcached-0\" (UID: \"90528e4b-0537-4477-b7d1-603c21bc9358\") " pod="openstack/memcached-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.149575 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpn7t\" (UniqueName: \"kubernetes.io/projected/90528e4b-0537-4477-b7d1-603c21bc9358-kube-api-access-rpn7t\") pod \"memcached-0\" (UID: \"90528e4b-0537-4477-b7d1-603c21bc9358\") " pod="openstack/memcached-0" Feb 25 13:36:10 crc kubenswrapper[4815]: I0225 13:36:10.338786 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 25 13:36:12 crc kubenswrapper[4815]: I0225 13:36:12.494849 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 13:36:12 crc kubenswrapper[4815]: I0225 13:36:12.495981 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 13:36:12 crc kubenswrapper[4815]: I0225 13:36:12.501564 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-zh7nt" Feb 25 13:36:12 crc kubenswrapper[4815]: I0225 13:36:12.507735 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 13:36:12 crc kubenswrapper[4815]: I0225 13:36:12.669740 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2phkz\" (UniqueName: \"kubernetes.io/projected/9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312-kube-api-access-2phkz\") pod \"kube-state-metrics-0\" (UID: \"9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312\") " pod="openstack/kube-state-metrics-0" Feb 25 13:36:12 crc kubenswrapper[4815]: I0225 13:36:12.771922 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2phkz\" (UniqueName: \"kubernetes.io/projected/9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312-kube-api-access-2phkz\") pod \"kube-state-metrics-0\" (UID: \"9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312\") " pod="openstack/kube-state-metrics-0" Feb 25 13:36:12 crc kubenswrapper[4815]: I0225 13:36:12.795577 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2phkz\" (UniqueName: \"kubernetes.io/projected/9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312-kube-api-access-2phkz\") pod \"kube-state-metrics-0\" (UID: \"9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312\") " pod="openstack/kube-state-metrics-0" Feb 25 13:36:12 crc kubenswrapper[4815]: I0225 13:36:12.820560 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 13:36:15 crc kubenswrapper[4815]: I0225 13:36:15.937320 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tjrr6"] Feb 25 13:36:15 crc kubenswrapper[4815]: I0225 13:36:15.939234 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:15 crc kubenswrapper[4815]: I0225 13:36:15.945593 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 25 13:36:15 crc kubenswrapper[4815]: I0225 13:36:15.946078 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-pw2ck" Feb 25 13:36:15 crc kubenswrapper[4815]: I0225 13:36:15.946785 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 25 13:36:15 crc kubenswrapper[4815]: I0225 13:36:15.949025 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-z74w4"] Feb 25 13:36:15 crc kubenswrapper[4815]: I0225 13:36:15.951435 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:15 crc kubenswrapper[4815]: I0225 13:36:15.958919 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tjrr6"] Feb 25 13:36:15 crc kubenswrapper[4815]: I0225 13:36:15.981047 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-z74w4"] Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.034426 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d08116-2a87-4ae4-8485-dab2bb05de4c-ovn-controller-tls-certs\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.034537 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dk4w\" (UniqueName: \"kubernetes.io/projected/d9d08116-2a87-4ae4-8485-dab2bb05de4c-kube-api-access-2dk4w\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.034572 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d9d08116-2a87-4ae4-8485-dab2bb05de4c-var-run-ovn\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.034615 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d9d08116-2a87-4ae4-8485-dab2bb05de4c-var-log-ovn\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.034801 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-scripts\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.034894 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d9d08116-2a87-4ae4-8485-dab2bb05de4c-scripts\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.034920 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d9d08116-2a87-4ae4-8485-dab2bb05de4c-var-run\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.034938 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9d08116-2a87-4ae4-8485-dab2bb05de4c-combined-ca-bundle\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.034981 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-var-log\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.035007 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-var-lib\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.035037 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-etc-ovs\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.035076 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjct9\" (UniqueName: \"kubernetes.io/projected/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-kube-api-access-xjct9\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.035122 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-var-run\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.136439 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-etc-ovs\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.136523 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjct9\" (UniqueName: \"kubernetes.io/projected/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-kube-api-access-xjct9\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.136582 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-var-run\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.136614 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d08116-2a87-4ae4-8485-dab2bb05de4c-ovn-controller-tls-certs\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.136652 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dk4w\" (UniqueName: \"kubernetes.io/projected/d9d08116-2a87-4ae4-8485-dab2bb05de4c-kube-api-access-2dk4w\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.136679 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d9d08116-2a87-4ae4-8485-dab2bb05de4c-var-run-ovn\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.136708 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d9d08116-2a87-4ae4-8485-dab2bb05de4c-var-log-ovn\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.136787 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-scripts\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.136862 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d9d08116-2a87-4ae4-8485-dab2bb05de4c-scripts\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.137597 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-etc-ovs\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.139471 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-var-run\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.139574 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d9d08116-2a87-4ae4-8485-dab2bb05de4c-var-run-ovn\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.139612 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d9d08116-2a87-4ae4-8485-dab2bb05de4c-var-run\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.140020 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d9d08116-2a87-4ae4-8485-dab2bb05de4c-scripts\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.140371 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d9d08116-2a87-4ae4-8485-dab2bb05de4c-var-run\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.140426 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9d08116-2a87-4ae4-8485-dab2bb05de4c-combined-ca-bundle\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.140564 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-var-log\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.140852 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-var-lib\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.149784 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d9d08116-2a87-4ae4-8485-dab2bb05de4c-var-log-ovn\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.152086 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-var-log\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.152302 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-scripts\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.153261 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjct9\" (UniqueName: \"kubernetes.io/projected/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-kube-api-access-xjct9\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.156524 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9d08116-2a87-4ae4-8485-dab2bb05de4c-ovn-controller-tls-certs\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.156681 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/67ea29a2-f5eb-41c5-84aa-e1ac90ab567b-var-lib\") pod \"ovn-controller-ovs-z74w4\" (UID: \"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b\") " pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.162176 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dk4w\" (UniqueName: \"kubernetes.io/projected/d9d08116-2a87-4ae4-8485-dab2bb05de4c-kube-api-access-2dk4w\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.162368 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9d08116-2a87-4ae4-8485-dab2bb05de4c-combined-ca-bundle\") pod \"ovn-controller-tjrr6\" (UID: \"d9d08116-2a87-4ae4-8485-dab2bb05de4c\") " pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.258346 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:16 crc kubenswrapper[4815]: I0225 13:36:16.281191 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:18 crc kubenswrapper[4815]: I0225 13:36:18.836554 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 25 13:36:18 crc kubenswrapper[4815]: I0225 13:36:18.838308 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:18 crc kubenswrapper[4815]: I0225 13:36:18.840097 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 25 13:36:18 crc kubenswrapper[4815]: I0225 13:36:18.844331 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 25 13:36:18 crc kubenswrapper[4815]: I0225 13:36:18.844421 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-hfjhs" Feb 25 13:36:18 crc kubenswrapper[4815]: I0225 13:36:18.845115 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 25 13:36:18 crc kubenswrapper[4815]: I0225 13:36:18.845150 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 25 13:36:18 crc kubenswrapper[4815]: I0225 13:36:18.849987 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 25 13:36:18 crc kubenswrapper[4815]: I0225 13:36:18.995568 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/205d03ae-ef40-447e-9176-eb6f6a6d139b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:18 crc kubenswrapper[4815]: I0225 13:36:18.995627 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/205d03ae-ef40-447e-9176-eb6f6a6d139b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:18 crc kubenswrapper[4815]: I0225 13:36:18.995667 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/205d03ae-ef40-447e-9176-eb6f6a6d139b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:18 crc kubenswrapper[4815]: I0225 13:36:18.995686 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf5mr\" (UniqueName: \"kubernetes.io/projected/205d03ae-ef40-447e-9176-eb6f6a6d139b-kube-api-access-wf5mr\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:18 crc kubenswrapper[4815]: I0225 13:36:18.995702 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/205d03ae-ef40-447e-9176-eb6f6a6d139b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:18 crc kubenswrapper[4815]: I0225 13:36:18.995767 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:18 crc kubenswrapper[4815]: I0225 13:36:18.995799 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/205d03ae-ef40-447e-9176-eb6f6a6d139b-config\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:18 crc kubenswrapper[4815]: I0225 13:36:18.995832 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/205d03ae-ef40-447e-9176-eb6f6a6d139b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.046661 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.049542 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.051969 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.052087 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.052376 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.052523 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-qjg5p" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.060419 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.097733 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.097795 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/205d03ae-ef40-447e-9176-eb6f6a6d139b-config\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.097834 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/205d03ae-ef40-447e-9176-eb6f6a6d139b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.097868 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/205d03ae-ef40-447e-9176-eb6f6a6d139b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.097918 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/205d03ae-ef40-447e-9176-eb6f6a6d139b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.097965 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/205d03ae-ef40-447e-9176-eb6f6a6d139b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.097984 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf5mr\" (UniqueName: \"kubernetes.io/projected/205d03ae-ef40-447e-9176-eb6f6a6d139b-kube-api-access-wf5mr\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.098005 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/205d03ae-ef40-447e-9176-eb6f6a6d139b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.098085 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.099556 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/205d03ae-ef40-447e-9176-eb6f6a6d139b-config\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.099676 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/205d03ae-ef40-447e-9176-eb6f6a6d139b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.100099 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/205d03ae-ef40-447e-9176-eb6f6a6d139b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.103950 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/205d03ae-ef40-447e-9176-eb6f6a6d139b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.103964 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/205d03ae-ef40-447e-9176-eb6f6a6d139b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.104294 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/205d03ae-ef40-447e-9176-eb6f6a6d139b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.116334 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf5mr\" (UniqueName: \"kubernetes.io/projected/205d03ae-ef40-447e-9176-eb6f6a6d139b-kube-api-access-wf5mr\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.125137 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"205d03ae-ef40-447e-9176-eb6f6a6d139b\") " pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.168717 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.199127 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/16d450f1-ecbb-4266-9818-afb1470b3a71-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.199182 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/16d450f1-ecbb-4266-9818-afb1470b3a71-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.199227 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16d450f1-ecbb-4266-9818-afb1470b3a71-config\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.199284 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16d450f1-ecbb-4266-9818-afb1470b3a71-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.199312 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7np66\" (UniqueName: \"kubernetes.io/projected/16d450f1-ecbb-4266-9818-afb1470b3a71-kube-api-access-7np66\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.199366 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.199398 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16d450f1-ecbb-4266-9818-afb1470b3a71-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.199437 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/16d450f1-ecbb-4266-9818-afb1470b3a71-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.301404 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16d450f1-ecbb-4266-9818-afb1470b3a71-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.301473 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7np66\" (UniqueName: \"kubernetes.io/projected/16d450f1-ecbb-4266-9818-afb1470b3a71-kube-api-access-7np66\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.301570 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.301620 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16d450f1-ecbb-4266-9818-afb1470b3a71-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.301656 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/16d450f1-ecbb-4266-9818-afb1470b3a71-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.301743 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/16d450f1-ecbb-4266-9818-afb1470b3a71-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.301787 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/16d450f1-ecbb-4266-9818-afb1470b3a71-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.302351 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16d450f1-ecbb-4266-9818-afb1470b3a71-config\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.301980 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.303016 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/16d450f1-ecbb-4266-9818-afb1470b3a71-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.303792 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16d450f1-ecbb-4266-9818-afb1470b3a71-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.304083 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16d450f1-ecbb-4266-9818-afb1470b3a71-config\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.315816 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/16d450f1-ecbb-4266-9818-afb1470b3a71-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.316761 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16d450f1-ecbb-4266-9818-afb1470b3a71-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.322262 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/16d450f1-ecbb-4266-9818-afb1470b3a71-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.326346 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7np66\" (UniqueName: \"kubernetes.io/projected/16d450f1-ecbb-4266-9818-afb1470b3a71-kube-api-access-7np66\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.339286 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"16d450f1-ecbb-4266-9818-afb1470b3a71\") " pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:19 crc kubenswrapper[4815]: I0225 13:36:19.365572 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:21 crc kubenswrapper[4815]: I0225 13:36:21.342107 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:36:21 crc kubenswrapper[4815]: I0225 13:36:21.343992 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:36:21 crc kubenswrapper[4815]: E0225 13:36:21.494459 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Feb 25 13:36:21 crc kubenswrapper[4815]: E0225 13:36:21.494678 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zsfhb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(4a91fa1d-2609-4b30-9ea1-8a5b10145c8d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 13:36:21 crc kubenswrapper[4815]: E0225 13:36:21.495920 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" Feb 25 13:36:22 crc kubenswrapper[4815]: E0225 13:36:22.122637 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" Feb 25 13:36:22 crc kubenswrapper[4815]: E0225 13:36:22.168366 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 25 13:36:22 crc kubenswrapper[4815]: E0225 13:36:22.168569 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mh528,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-74q86_openstack(11720769-5e77-4722-bfba-f208fcc3276d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 13:36:22 crc kubenswrapper[4815]: E0225 13:36:22.169859 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-74q86" podUID="11720769-5e77-4722-bfba-f208fcc3276d" Feb 25 13:36:22 crc kubenswrapper[4815]: E0225 13:36:22.216050 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 25 13:36:22 crc kubenswrapper[4815]: E0225 13:36:22.216211 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tp8ns,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-bmkxb_openstack(9bb1d381-cd64-457f-b81a-49851aae7c68): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 13:36:22 crc kubenswrapper[4815]: E0225 13:36:22.217434 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-bmkxb" podUID="9bb1d381-cd64-457f-b81a-49851aae7c68" Feb 25 13:36:22 crc kubenswrapper[4815]: E0225 13:36:22.256682 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 25 13:36:22 crc kubenswrapper[4815]: E0225 13:36:22.256858 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lwtwf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-ktjsd_openstack(9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 13:36:22 crc kubenswrapper[4815]: E0225 13:36:22.258050 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-ktjsd" podUID="9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5" Feb 25 13:36:22 crc kubenswrapper[4815]: E0225 13:36:22.304908 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 25 13:36:22 crc kubenswrapper[4815]: E0225 13:36:22.305363 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-86dqs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-npxbr_openstack(81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 13:36:22 crc kubenswrapper[4815]: E0225 13:36:22.306768 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-npxbr" podUID="81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6" Feb 25 13:36:22 crc kubenswrapper[4815]: W0225 13:36:22.775412 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52d1a204_9bb6_40e1_a63c_09ffe6016cd9.slice/crio-1b745b3965cb968a4e3d63b0c1b0fbb5bfbfbfe6f024e4e55242b70a074d09ca WatchSource:0}: Error finding container 1b745b3965cb968a4e3d63b0c1b0fbb5bfbfbfe6f024e4e55242b70a074d09ca: Status 404 returned error can't find the container with id 1b745b3965cb968a4e3d63b0c1b0fbb5bfbfbfe6f024e4e55242b70a074d09ca Feb 25 13:36:22 crc kubenswrapper[4815]: I0225 13:36:22.802760 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 25 13:36:22 crc kubenswrapper[4815]: I0225 13:36:22.810697 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 25 13:36:22 crc kubenswrapper[4815]: I0225 13:36:22.818107 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 25 13:36:22 crc kubenswrapper[4815]: I0225 13:36:22.869936 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 25 13:36:22 crc kubenswrapper[4815]: I0225 13:36:22.928193 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 13:36:22 crc kubenswrapper[4815]: I0225 13:36:22.945816 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tjrr6"] Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.007608 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-z74w4"] Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.114791 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tjrr6" event={"ID":"d9d08116-2a87-4ae4-8485-dab2bb05de4c","Type":"ContainerStarted","Data":"85c1eb16970207452142b739decf33a403a0704f9ea8bfaebc2e09a6da6f657d"} Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.116093 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-z74w4" event={"ID":"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b","Type":"ContainerStarted","Data":"29cd7d83f24ace6e93fa313a8dc861e38d5ccdb5c57998a1b53f43ceafc6720e"} Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.117300 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850","Type":"ContainerStarted","Data":"dd09400c2c97a289039dc7d19e857ca38cb6d004bb19a82fda9d8464fd9c23c8"} Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.118257 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"90528e4b-0537-4477-b7d1-603c21bc9358","Type":"ContainerStarted","Data":"23b06c73eef5a1533c16f0ee2ceb4bcf226ba0ad2e2973e7e25995a7020fa7be"} Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.119782 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"16d450f1-ecbb-4266-9818-afb1470b3a71","Type":"ContainerStarted","Data":"db2e4f9adcc30d590e1ce2d945f79bf6376788e75ee50e14303909df2cd16c6d"} Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.120792 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312","Type":"ContainerStarted","Data":"05da5e98867f3a1239d0b8bca687c4a5e4cebc0a0f2e56aac9e1f9c81e25c716"} Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.122563 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"52d1a204-9bb6-40e1-a63c-09ffe6016cd9","Type":"ContainerStarted","Data":"1b745b3965cb968a4e3d63b0c1b0fbb5bfbfbfe6f024e4e55242b70a074d09ca"} Feb 25 13:36:23 crc kubenswrapper[4815]: E0225 13:36:23.124786 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-bmkxb" podUID="9bb1d381-cd64-457f-b81a-49851aae7c68" Feb 25 13:36:23 crc kubenswrapper[4815]: E0225 13:36:23.124897 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-ktjsd" podUID="9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5" Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.468339 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-74q86" Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.473397 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-npxbr" Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.575310 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11720769-5e77-4722-bfba-f208fcc3276d-config\") pod \"11720769-5e77-4722-bfba-f208fcc3276d\" (UID: \"11720769-5e77-4722-bfba-f208fcc3276d\") " Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.575363 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mh528\" (UniqueName: \"kubernetes.io/projected/11720769-5e77-4722-bfba-f208fcc3276d-kube-api-access-mh528\") pod \"11720769-5e77-4722-bfba-f208fcc3276d\" (UID: \"11720769-5e77-4722-bfba-f208fcc3276d\") " Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.575402 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-config\") pod \"81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6\" (UID: \"81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6\") " Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.575430 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-dns-svc\") pod \"81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6\" (UID: \"81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6\") " Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.575561 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86dqs\" (UniqueName: \"kubernetes.io/projected/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-kube-api-access-86dqs\") pod \"81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6\" (UID: \"81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6\") " Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.576791 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11720769-5e77-4722-bfba-f208fcc3276d-config" (OuterVolumeSpecName: "config") pod "11720769-5e77-4722-bfba-f208fcc3276d" (UID: "11720769-5e77-4722-bfba-f208fcc3276d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.577315 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6" (UID: "81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.577462 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-config" (OuterVolumeSpecName: "config") pod "81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6" (UID: "81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.591357 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11720769-5e77-4722-bfba-f208fcc3276d-kube-api-access-mh528" (OuterVolumeSpecName: "kube-api-access-mh528") pod "11720769-5e77-4722-bfba-f208fcc3276d" (UID: "11720769-5e77-4722-bfba-f208fcc3276d"). InnerVolumeSpecName "kube-api-access-mh528". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.596642 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-kube-api-access-86dqs" (OuterVolumeSpecName: "kube-api-access-86dqs") pod "81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6" (UID: "81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6"). InnerVolumeSpecName "kube-api-access-86dqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.677233 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86dqs\" (UniqueName: \"kubernetes.io/projected/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-kube-api-access-86dqs\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.677266 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11720769-5e77-4722-bfba-f208fcc3276d-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.677275 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mh528\" (UniqueName: \"kubernetes.io/projected/11720769-5e77-4722-bfba-f208fcc3276d-kube-api-access-mh528\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.677284 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.677293 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:23 crc kubenswrapper[4815]: I0225 13:36:23.936231 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 25 13:36:24 crc kubenswrapper[4815]: I0225 13:36:24.129746 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-npxbr" event={"ID":"81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6","Type":"ContainerDied","Data":"e64591df5903b5ddf98ae24578aea53c1f5957213ad4fec14b34ad5d24e05c10"} Feb 25 13:36:24 crc kubenswrapper[4815]: I0225 13:36:24.130151 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-npxbr" Feb 25 13:36:24 crc kubenswrapper[4815]: I0225 13:36:24.136129 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0b98e7b2-3068-4f59-b25f-4755bd752b49","Type":"ContainerStarted","Data":"960d7fcb921daf08a33279bdf194bd3ec4837f2737f63bcd0cb1c55a13429919"} Feb 25 13:36:24 crc kubenswrapper[4815]: I0225 13:36:24.138268 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-74q86" event={"ID":"11720769-5e77-4722-bfba-f208fcc3276d","Type":"ContainerDied","Data":"8b27f1495338a9f8d2297103bbcae67f651a485118e4544bce4a7025aafcdec8"} Feb 25 13:36:24 crc kubenswrapper[4815]: I0225 13:36:24.138978 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-74q86" Feb 25 13:36:24 crc kubenswrapper[4815]: I0225 13:36:24.198983 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-npxbr"] Feb 25 13:36:24 crc kubenswrapper[4815]: I0225 13:36:24.206738 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-npxbr"] Feb 25 13:36:24 crc kubenswrapper[4815]: I0225 13:36:24.229356 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-74q86"] Feb 25 13:36:24 crc kubenswrapper[4815]: I0225 13:36:24.236454 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-74q86"] Feb 25 13:36:24 crc kubenswrapper[4815]: W0225 13:36:24.313889 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod205d03ae_ef40_447e_9176_eb6f6a6d139b.slice/crio-591ea2cccd101bad93b444f47b8c6fa6ff2a854f77d262a08134526d4efbfd77 WatchSource:0}: Error finding container 591ea2cccd101bad93b444f47b8c6fa6ff2a854f77d262a08134526d4efbfd77: Status 404 returned error can't find the container with id 591ea2cccd101bad93b444f47b8c6fa6ff2a854f77d262a08134526d4efbfd77 Feb 25 13:36:24 crc kubenswrapper[4815]: I0225 13:36:24.945806 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11720769-5e77-4722-bfba-f208fcc3276d" path="/var/lib/kubelet/pods/11720769-5e77-4722-bfba-f208fcc3276d/volumes" Feb 25 13:36:24 crc kubenswrapper[4815]: I0225 13:36:24.946210 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6" path="/var/lib/kubelet/pods/81e4eb3f-c2c2-4b39-94d0-612a3b7b2bb6/volumes" Feb 25 13:36:25 crc kubenswrapper[4815]: I0225 13:36:25.146533 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"205d03ae-ef40-447e-9176-eb6f6a6d139b","Type":"ContainerStarted","Data":"591ea2cccd101bad93b444f47b8c6fa6ff2a854f77d262a08134526d4efbfd77"} Feb 25 13:36:30 crc kubenswrapper[4815]: I0225 13:36:30.181720 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"205d03ae-ef40-447e-9176-eb6f6a6d139b","Type":"ContainerStarted","Data":"1c0fb0a5ef5cd2b8ba1dab2df821f4d8307329c3bd6d0b73ab8380a3acbe5a36"} Feb 25 13:36:30 crc kubenswrapper[4815]: I0225 13:36:30.183336 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850","Type":"ContainerStarted","Data":"3076462a510972042b90d4d6229977452d4e4873eac58e0a64e21dfc5aaf2473"} Feb 25 13:36:30 crc kubenswrapper[4815]: I0225 13:36:30.184542 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"90528e4b-0537-4477-b7d1-603c21bc9358","Type":"ContainerStarted","Data":"8d54963cc0a6745fb9589b19fb53fc45535d8cf8a6c87cf4e8cdcce321b5f9ef"} Feb 25 13:36:30 crc kubenswrapper[4815]: I0225 13:36:30.184679 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 25 13:36:30 crc kubenswrapper[4815]: I0225 13:36:30.185973 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"16d450f1-ecbb-4266-9818-afb1470b3a71","Type":"ContainerStarted","Data":"b9f8521b22b9eccf01690f20f44a389fd42b4126504ac668665abecd15ff9a67"} Feb 25 13:36:30 crc kubenswrapper[4815]: I0225 13:36:30.187737 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312","Type":"ContainerStarted","Data":"e7e1d82b05c4847e358e488d560ac8a8df154f0e40cfd12e322ad254ac268410"} Feb 25 13:36:30 crc kubenswrapper[4815]: I0225 13:36:30.187886 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 25 13:36:30 crc kubenswrapper[4815]: I0225 13:36:30.189159 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"52d1a204-9bb6-40e1-a63c-09ffe6016cd9","Type":"ContainerStarted","Data":"f5c8339f6761db9af87f6727653bb0248a872d06ecb5be2a9fb0365e706bcf34"} Feb 25 13:36:30 crc kubenswrapper[4815]: I0225 13:36:30.190963 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tjrr6" event={"ID":"d9d08116-2a87-4ae4-8485-dab2bb05de4c","Type":"ContainerStarted","Data":"354d17e292310f081ec0c20c338ddac88a45705d344d803ce1167ca5d96ade9d"} Feb 25 13:36:30 crc kubenswrapper[4815]: I0225 13:36:30.191081 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-tjrr6" Feb 25 13:36:30 crc kubenswrapper[4815]: I0225 13:36:30.192086 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-z74w4" event={"ID":"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b","Type":"ContainerStarted","Data":"ce8728c3ec6fe67ab29c757f440b8f9537ea497eb3c25c7b65d0b552c5a872fc"} Feb 25 13:36:30 crc kubenswrapper[4815]: I0225 13:36:30.228267 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=11.426019154 podStartE2EDuration="18.228252319s" podCreationTimestamp="2026-02-25 13:36:12 +0000 UTC" firstStartedPulling="2026-02-25 13:36:22.926578597 +0000 UTC m=+940.727676651" lastFinishedPulling="2026-02-25 13:36:29.728811752 +0000 UTC m=+947.529909816" observedRunningTime="2026-02-25 13:36:30.223795121 +0000 UTC m=+948.024893175" watchObservedRunningTime="2026-02-25 13:36:30.228252319 +0000 UTC m=+948.029350373" Feb 25 13:36:30 crc kubenswrapper[4815]: I0225 13:36:30.243941 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=15.124130749 podStartE2EDuration="21.243917913s" podCreationTimestamp="2026-02-25 13:36:09 +0000 UTC" firstStartedPulling="2026-02-25 13:36:22.795830385 +0000 UTC m=+940.596928439" lastFinishedPulling="2026-02-25 13:36:28.915617549 +0000 UTC m=+946.716715603" observedRunningTime="2026-02-25 13:36:30.241487268 +0000 UTC m=+948.042585332" watchObservedRunningTime="2026-02-25 13:36:30.243917913 +0000 UTC m=+948.045015967" Feb 25 13:36:30 crc kubenswrapper[4815]: I0225 13:36:30.292230 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-tjrr6" podStartSLOduration=9.223947825 podStartE2EDuration="15.292212196s" podCreationTimestamp="2026-02-25 13:36:15 +0000 UTC" firstStartedPulling="2026-02-25 13:36:22.934757479 +0000 UTC m=+940.735855533" lastFinishedPulling="2026-02-25 13:36:29.00302185 +0000 UTC m=+946.804119904" observedRunningTime="2026-02-25 13:36:30.288930114 +0000 UTC m=+948.090028168" watchObservedRunningTime="2026-02-25 13:36:30.292212196 +0000 UTC m=+948.093310250" Feb 25 13:36:31 crc kubenswrapper[4815]: I0225 13:36:31.201703 4815 generic.go:334] "Generic (PLEG): container finished" podID="67ea29a2-f5eb-41c5-84aa-e1ac90ab567b" containerID="ce8728c3ec6fe67ab29c757f440b8f9537ea497eb3c25c7b65d0b552c5a872fc" exitCode=0 Feb 25 13:36:31 crc kubenswrapper[4815]: I0225 13:36:31.201764 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-z74w4" event={"ID":"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b","Type":"ContainerDied","Data":"ce8728c3ec6fe67ab29c757f440b8f9537ea497eb3c25c7b65d0b552c5a872fc"} Feb 25 13:36:32 crc kubenswrapper[4815]: I0225 13:36:32.210414 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"16d450f1-ecbb-4266-9818-afb1470b3a71","Type":"ContainerStarted","Data":"db27cd85e9dd800190e7675bcd0b55d0b669e60e1f60a5a293bfe24f3fbda831"} Feb 25 13:36:32 crc kubenswrapper[4815]: I0225 13:36:32.213850 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-z74w4" event={"ID":"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b","Type":"ContainerStarted","Data":"96c70075e0dcd145ef7339ca7857cccbe3e9b63d30a615d29822558b10beb294"} Feb 25 13:36:32 crc kubenswrapper[4815]: I0225 13:36:32.213897 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-z74w4" event={"ID":"67ea29a2-f5eb-41c5-84aa-e1ac90ab567b","Type":"ContainerStarted","Data":"e89d1c9ad0697e2900613f1aa91da1683862c866c48c21a627064d94293e11be"} Feb 25 13:36:32 crc kubenswrapper[4815]: I0225 13:36:32.213976 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:32 crc kubenswrapper[4815]: I0225 13:36:32.214012 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:36:32 crc kubenswrapper[4815]: I0225 13:36:32.215663 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"205d03ae-ef40-447e-9176-eb6f6a6d139b","Type":"ContainerStarted","Data":"c18ca9438c31cf7a12ed2e416ee434b7250dbc11b447b4a2a25930dd586b1df0"} Feb 25 13:36:32 crc kubenswrapper[4815]: I0225 13:36:32.233681 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=5.637451777 podStartE2EDuration="14.233667049s" podCreationTimestamp="2026-02-25 13:36:18 +0000 UTC" firstStartedPulling="2026-02-25 13:36:22.878386027 +0000 UTC m=+940.679484081" lastFinishedPulling="2026-02-25 13:36:31.474601299 +0000 UTC m=+949.275699353" observedRunningTime="2026-02-25 13:36:32.23110172 +0000 UTC m=+950.032199794" watchObservedRunningTime="2026-02-25 13:36:32.233667049 +0000 UTC m=+950.034765103" Feb 25 13:36:32 crc kubenswrapper[4815]: I0225 13:36:32.256790 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=8.088579706 podStartE2EDuration="15.256774254s" podCreationTimestamp="2026-02-25 13:36:17 +0000 UTC" firstStartedPulling="2026-02-25 13:36:24.315862394 +0000 UTC m=+942.116960438" lastFinishedPulling="2026-02-25 13:36:31.484056892 +0000 UTC m=+949.285154986" observedRunningTime="2026-02-25 13:36:32.247668922 +0000 UTC m=+950.048766976" watchObservedRunningTime="2026-02-25 13:36:32.256774254 +0000 UTC m=+950.057872308" Feb 25 13:36:32 crc kubenswrapper[4815]: I0225 13:36:32.274612 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-z74w4" podStartSLOduration=11.34823755 podStartE2EDuration="17.274596045s" podCreationTimestamp="2026-02-25 13:36:15 +0000 UTC" firstStartedPulling="2026-02-25 13:36:22.999415898 +0000 UTC m=+940.800513982" lastFinishedPulling="2026-02-25 13:36:28.925774423 +0000 UTC m=+946.726872477" observedRunningTime="2026-02-25 13:36:32.269132986 +0000 UTC m=+950.070231040" watchObservedRunningTime="2026-02-25 13:36:32.274596045 +0000 UTC m=+950.075694099" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.169349 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.170182 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.226569 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.246136 4815 generic.go:334] "Generic (PLEG): container finished" podID="52d1a204-9bb6-40e1-a63c-09ffe6016cd9" containerID="f5c8339f6761db9af87f6727653bb0248a872d06ecb5be2a9fb0365e706bcf34" exitCode=0 Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.246219 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"52d1a204-9bb6-40e1-a63c-09ffe6016cd9","Type":"ContainerDied","Data":"f5c8339f6761db9af87f6727653bb0248a872d06ecb5be2a9fb0365e706bcf34"} Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.251193 4815 generic.go:334] "Generic (PLEG): container finished" podID="c96edb5d-f9a9-4dc1-a9c2-e3def3efb850" containerID="3076462a510972042b90d4d6229977452d4e4873eac58e0a64e21dfc5aaf2473" exitCode=0 Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.251814 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850","Type":"ContainerDied","Data":"3076462a510972042b90d4d6229977452d4e4873eac58e0a64e21dfc5aaf2473"} Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.314145 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.367206 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.370624 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.413066 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.546613 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ktjsd"] Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.574753 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-74nkx"] Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.584208 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.602204 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-v2kjv"] Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.603133 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.604585 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.604972 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.616414 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-74nkx"] Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.631685 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-v2kjv"] Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.667268 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.667308 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-combined-ca-bundle\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.667365 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-ovs-rundir\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.667383 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-74nkx\" (UID: \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\") " pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.667416 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-74nkx\" (UID: \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\") " pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.667444 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-config\") pod \"dnsmasq-dns-5bf47b49b7-74nkx\" (UID: \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\") " pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.667468 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-config\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.667483 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-ovn-rundir\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.667522 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25b28\" (UniqueName: \"kubernetes.io/projected/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-kube-api-access-25b28\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.667546 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tknfg\" (UniqueName: \"kubernetes.io/projected/5926509f-0818-4ec5-96d3-ee2e9cb1504f-kube-api-access-tknfg\") pod \"dnsmasq-dns-5bf47b49b7-74nkx\" (UID: \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\") " pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.773620 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-config\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.769500 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-config\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.773742 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-ovn-rundir\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.773958 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-ovn-rundir\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.774074 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25b28\" (UniqueName: \"kubernetes.io/projected/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-kube-api-access-25b28\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.774548 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tknfg\" (UniqueName: \"kubernetes.io/projected/5926509f-0818-4ec5-96d3-ee2e9cb1504f-kube-api-access-tknfg\") pod \"dnsmasq-dns-5bf47b49b7-74nkx\" (UID: \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\") " pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.774973 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.777597 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-combined-ca-bundle\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.780574 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-ovs-rundir\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.788440 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-combined-ca-bundle\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.788912 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.788960 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-ovs-rundir\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.789020 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-74nkx\" (UID: \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\") " pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.789828 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-74nkx\" (UID: \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\") " pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.789939 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-74nkx\" (UID: \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\") " pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.790011 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-config\") pod \"dnsmasq-dns-5bf47b49b7-74nkx\" (UID: \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\") " pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.791379 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-74nkx\" (UID: \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\") " pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.791706 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-config\") pod \"dnsmasq-dns-5bf47b49b7-74nkx\" (UID: \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\") " pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.793373 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25b28\" (UniqueName: \"kubernetes.io/projected/8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd-kube-api-access-25b28\") pod \"ovn-controller-metrics-v2kjv\" (UID: \"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd\") " pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.810185 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tknfg\" (UniqueName: \"kubernetes.io/projected/5926509f-0818-4ec5-96d3-ee2e9cb1504f-kube-api-access-tknfg\") pod \"dnsmasq-dns-5bf47b49b7-74nkx\" (UID: \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\") " pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.848131 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-ktjsd" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.901626 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmkxb"] Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.957408 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.959918 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-v2kjv" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.992227 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-dns-svc\") pod \"9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5\" (UID: \"9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5\") " Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.992360 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-config\") pod \"9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5\" (UID: \"9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5\") " Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.992456 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwtwf\" (UniqueName: \"kubernetes.io/projected/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-kube-api-access-lwtwf\") pod \"9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5\" (UID: \"9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5\") " Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.992761 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5" (UID: "9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.993103 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-config" (OuterVolumeSpecName: "config") pod "9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5" (UID: "9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:34 crc kubenswrapper[4815]: I0225 13:36:34.995202 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-kube-api-access-lwtwf" (OuterVolumeSpecName: "kube-api-access-lwtwf") pod "9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5" (UID: "9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5"). InnerVolumeSpecName "kube-api-access-lwtwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.001712 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-hktjt"] Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.003174 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-hktjt"] Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.003278 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.009891 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.094660 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-hktjt\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.094900 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-dns-svc\") pod \"dnsmasq-dns-8554648995-hktjt\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.095021 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-config\") pod \"dnsmasq-dns-8554648995-hktjt\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.095058 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2846\" (UniqueName: \"kubernetes.io/projected/4f6d715d-8d19-48be-b617-7bf2ea9b6098-kube-api-access-w2846\") pod \"dnsmasq-dns-8554648995-hktjt\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.095084 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-hktjt\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.095221 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.095233 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwtwf\" (UniqueName: \"kubernetes.io/projected/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-kube-api-access-lwtwf\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.095243 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.196347 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-dns-svc\") pod \"dnsmasq-dns-8554648995-hktjt\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.196399 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-config\") pod \"dnsmasq-dns-8554648995-hktjt\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.196423 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2846\" (UniqueName: \"kubernetes.io/projected/4f6d715d-8d19-48be-b617-7bf2ea9b6098-kube-api-access-w2846\") pod \"dnsmasq-dns-8554648995-hktjt\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.196444 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-hktjt\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.196486 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-hktjt\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.197301 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-dns-svc\") pod \"dnsmasq-dns-8554648995-hktjt\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.197531 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-hktjt\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.197557 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-hktjt\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.197557 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-config\") pod \"dnsmasq-dns-8554648995-hktjt\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.213189 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2846\" (UniqueName: \"kubernetes.io/projected/4f6d715d-8d19-48be-b617-7bf2ea9b6098-kube-api-access-w2846\") pod \"dnsmasq-dns-8554648995-hktjt\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.225087 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bmkxb" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.261056 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c96edb5d-f9a9-4dc1-a9c2-e3def3efb850","Type":"ContainerStarted","Data":"aaa286dcfa26defa244e91ab6834bad6f0d1384952ca86b9346a0998a632c866"} Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.262436 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"52d1a204-9bb6-40e1-a63c-09ffe6016cd9","Type":"ContainerStarted","Data":"b36a4893b893919927a7896a92c454b2ad886d80595000938bc1eb32122b7489"} Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.263664 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bmkxb" event={"ID":"9bb1d381-cd64-457f-b81a-49851aae7c68","Type":"ContainerDied","Data":"4835449f02f41a94f4c8339cfe1cc5fbe8c637d8313506c73a373e385e49219b"} Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.263716 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bmkxb" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.265444 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-ktjsd" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.269572 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-ktjsd" event={"ID":"9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5","Type":"ContainerDied","Data":"c424a1af5ed8fdede4017fd380fef7224338f87c27113a3c7b11ddfe9f2864cc"} Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.297470 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=20.905218538 podStartE2EDuration="27.297449962s" podCreationTimestamp="2026-02-25 13:36:08 +0000 UTC" firstStartedPulling="2026-02-25 13:36:22.796335501 +0000 UTC m=+940.597433555" lastFinishedPulling="2026-02-25 13:36:29.188566915 +0000 UTC m=+946.989664979" observedRunningTime="2026-02-25 13:36:35.29155461 +0000 UTC m=+953.092652654" watchObservedRunningTime="2026-02-25 13:36:35.297449962 +0000 UTC m=+953.098548016" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.299404 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp8ns\" (UniqueName: \"kubernetes.io/projected/9bb1d381-cd64-457f-b81a-49851aae7c68-kube-api-access-tp8ns\") pod \"9bb1d381-cd64-457f-b81a-49851aae7c68\" (UID: \"9bb1d381-cd64-457f-b81a-49851aae7c68\") " Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.299542 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb1d381-cd64-457f-b81a-49851aae7c68-config\") pod \"9bb1d381-cd64-457f-b81a-49851aae7c68\" (UID: \"9bb1d381-cd64-457f-b81a-49851aae7c68\") " Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.299610 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bb1d381-cd64-457f-b81a-49851aae7c68-dns-svc\") pod \"9bb1d381-cd64-457f-b81a-49851aae7c68\" (UID: \"9bb1d381-cd64-457f-b81a-49851aae7c68\") " Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.300349 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bb1d381-cd64-457f-b81a-49851aae7c68-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9bb1d381-cd64-457f-b81a-49851aae7c68" (UID: "9bb1d381-cd64-457f-b81a-49851aae7c68"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.304287 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bb1d381-cd64-457f-b81a-49851aae7c68-config" (OuterVolumeSpecName: "config") pod "9bb1d381-cd64-457f-b81a-49851aae7c68" (UID: "9bb1d381-cd64-457f-b81a-49851aae7c68"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.310759 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bb1d381-cd64-457f-b81a-49851aae7c68-kube-api-access-tp8ns" (OuterVolumeSpecName: "kube-api-access-tp8ns") pod "9bb1d381-cd64-457f-b81a-49851aae7c68" (UID: "9bb1d381-cd64-457f-b81a-49851aae7c68"). InnerVolumeSpecName "kube-api-access-tp8ns". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.331485 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-74nkx"] Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.331667 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.331797 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.340620 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.342683 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=22.035599918 podStartE2EDuration="28.342657839s" podCreationTimestamp="2026-02-25 13:36:07 +0000 UTC" firstStartedPulling="2026-02-25 13:36:22.778141389 +0000 UTC m=+940.579239443" lastFinishedPulling="2026-02-25 13:36:29.0851993 +0000 UTC m=+946.886297364" observedRunningTime="2026-02-25 13:36:35.331980529 +0000 UTC m=+953.133078583" watchObservedRunningTime="2026-02-25 13:36:35.342657839 +0000 UTC m=+953.143755913" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.407853 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp8ns\" (UniqueName: \"kubernetes.io/projected/9bb1d381-cd64-457f-b81a-49851aae7c68-kube-api-access-tp8ns\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.407886 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb1d381-cd64-457f-b81a-49851aae7c68-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.407896 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bb1d381-cd64-457f-b81a-49851aae7c68-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.501215 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ktjsd"] Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.527271 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-v2kjv"] Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.530295 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ktjsd"] Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.621786 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmkxb"] Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.628555 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmkxb"] Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.765109 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.766230 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.769397 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-f6kmr" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.773418 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.773628 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.773734 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.783825 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.816878 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ljdq\" (UniqueName: \"kubernetes.io/projected/ebc28857-40fb-4026-b7df-a2033618af75-kube-api-access-5ljdq\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.816948 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebc28857-40fb-4026-b7df-a2033618af75-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.817083 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebc28857-40fb-4026-b7df-a2033618af75-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.817132 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebc28857-40fb-4026-b7df-a2033618af75-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.817236 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebc28857-40fb-4026-b7df-a2033618af75-config\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.817290 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ebc28857-40fb-4026-b7df-a2033618af75-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.817312 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ebc28857-40fb-4026-b7df-a2033618af75-scripts\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.838920 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-hktjt"] Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.918595 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebc28857-40fb-4026-b7df-a2033618af75-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.918637 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebc28857-40fb-4026-b7df-a2033618af75-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.918709 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebc28857-40fb-4026-b7df-a2033618af75-config\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.918799 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ebc28857-40fb-4026-b7df-a2033618af75-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.918827 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ebc28857-40fb-4026-b7df-a2033618af75-scripts\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.918863 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ljdq\" (UniqueName: \"kubernetes.io/projected/ebc28857-40fb-4026-b7df-a2033618af75-kube-api-access-5ljdq\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.918891 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebc28857-40fb-4026-b7df-a2033618af75-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.920477 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ebc28857-40fb-4026-b7df-a2033618af75-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.920866 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebc28857-40fb-4026-b7df-a2033618af75-config\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.921084 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ebc28857-40fb-4026-b7df-a2033618af75-scripts\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.928842 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebc28857-40fb-4026-b7df-a2033618af75-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.930986 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebc28857-40fb-4026-b7df-a2033618af75-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: W0225 13:36:35.937053 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f6d715d_8d19_48be_b617_7bf2ea9b6098.slice/crio-cc6e372c493fe6ad5b38db578ed73ee78fdd863b0da5607ef9a6b9af08b53b20 WatchSource:0}: Error finding container cc6e372c493fe6ad5b38db578ed73ee78fdd863b0da5607ef9a6b9af08b53b20: Status 404 returned error can't find the container with id cc6e372c493fe6ad5b38db578ed73ee78fdd863b0da5607ef9a6b9af08b53b20 Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.940556 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ljdq\" (UniqueName: \"kubernetes.io/projected/ebc28857-40fb-4026-b7df-a2033618af75-kube-api-access-5ljdq\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:35 crc kubenswrapper[4815]: I0225 13:36:35.944779 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebc28857-40fb-4026-b7df-a2033618af75-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ebc28857-40fb-4026-b7df-a2033618af75\") " pod="openstack/ovn-northd-0" Feb 25 13:36:36 crc kubenswrapper[4815]: I0225 13:36:36.105215 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 25 13:36:36 crc kubenswrapper[4815]: I0225 13:36:36.297992 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d","Type":"ContainerStarted","Data":"e633733b61b772b04a5fcbcaa5076b5a7353c6b7fd094307d1bc64bbb9741d6c"} Feb 25 13:36:36 crc kubenswrapper[4815]: I0225 13:36:36.299699 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-hktjt" event={"ID":"4f6d715d-8d19-48be-b617-7bf2ea9b6098","Type":"ContainerStarted","Data":"cc6e372c493fe6ad5b38db578ed73ee78fdd863b0da5607ef9a6b9af08b53b20"} Feb 25 13:36:36 crc kubenswrapper[4815]: I0225 13:36:36.302679 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" event={"ID":"5926509f-0818-4ec5-96d3-ee2e9cb1504f","Type":"ContainerStarted","Data":"8c3dd5b64e6d21e8a1ef6438d5038832d8c2b167a70c5da5403d26cdcc959566"} Feb 25 13:36:36 crc kubenswrapper[4815]: I0225 13:36:36.302719 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" event={"ID":"5926509f-0818-4ec5-96d3-ee2e9cb1504f","Type":"ContainerStarted","Data":"fd3ebc7986cc63fc3ed05f1fc7ce820433f5b0dd15e0780f2138d24c2c5ecc07"} Feb 25 13:36:36 crc kubenswrapper[4815]: I0225 13:36:36.305991 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-v2kjv" event={"ID":"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd","Type":"ContainerStarted","Data":"19b504a24ec1859c9974bbcaed7c486442ede28e8fd8342290a82ab56cd84783"} Feb 25 13:36:36 crc kubenswrapper[4815]: I0225 13:36:36.306031 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-v2kjv" event={"ID":"8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd","Type":"ContainerStarted","Data":"ea771970582a63d83d336f3b2304a132cdaf49f765bc3182039397c73c0c65ad"} Feb 25 13:36:36 crc kubenswrapper[4815]: I0225 13:36:36.361836 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-v2kjv" podStartSLOduration=2.361818539 podStartE2EDuration="2.361818539s" podCreationTimestamp="2026-02-25 13:36:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:36:36.358220197 +0000 UTC m=+954.159318261" watchObservedRunningTime="2026-02-25 13:36:36.361818539 +0000 UTC m=+954.162916593" Feb 25 13:36:36 crc kubenswrapper[4815]: I0225 13:36:36.534364 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 25 13:36:36 crc kubenswrapper[4815]: I0225 13:36:36.955177 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5" path="/var/lib/kubelet/pods/9a2c8dfd-6f49-4a9c-b3ae-ea354a0da6f5/volumes" Feb 25 13:36:36 crc kubenswrapper[4815]: I0225 13:36:36.956305 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bb1d381-cd64-457f-b81a-49851aae7c68" path="/var/lib/kubelet/pods/9bb1d381-cd64-457f-b81a-49851aae7c68/volumes" Feb 25 13:36:37 crc kubenswrapper[4815]: I0225 13:36:37.329852 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ebc28857-40fb-4026-b7df-a2033618af75","Type":"ContainerStarted","Data":"8e50890783b140f27aedd5a993be82da59d25c25dfde566ca63e50f140eee001"} Feb 25 13:36:37 crc kubenswrapper[4815]: I0225 13:36:37.377288 4815 generic.go:334] "Generic (PLEG): container finished" podID="4f6d715d-8d19-48be-b617-7bf2ea9b6098" containerID="5cfd11297a7ab195f05dd312402f3acc2d4e7fbbf06cdddd2e6c5330680e0176" exitCode=0 Feb 25 13:36:37 crc kubenswrapper[4815]: I0225 13:36:37.377390 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-hktjt" event={"ID":"4f6d715d-8d19-48be-b617-7bf2ea9b6098","Type":"ContainerDied","Data":"5cfd11297a7ab195f05dd312402f3acc2d4e7fbbf06cdddd2e6c5330680e0176"} Feb 25 13:36:37 crc kubenswrapper[4815]: I0225 13:36:37.403045 4815 generic.go:334] "Generic (PLEG): container finished" podID="5926509f-0818-4ec5-96d3-ee2e9cb1504f" containerID="8c3dd5b64e6d21e8a1ef6438d5038832d8c2b167a70c5da5403d26cdcc959566" exitCode=0 Feb 25 13:36:37 crc kubenswrapper[4815]: I0225 13:36:37.404294 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" event={"ID":"5926509f-0818-4ec5-96d3-ee2e9cb1504f","Type":"ContainerDied","Data":"8c3dd5b64e6d21e8a1ef6438d5038832d8c2b167a70c5da5403d26cdcc959566"} Feb 25 13:36:38 crc kubenswrapper[4815]: I0225 13:36:38.419349 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ebc28857-40fb-4026-b7df-a2033618af75","Type":"ContainerStarted","Data":"4f7884d026b0e662ee8fcfe7e1cacadd8ee039ff973171b803969470eecdbc42"} Feb 25 13:36:38 crc kubenswrapper[4815]: I0225 13:36:38.419769 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ebc28857-40fb-4026-b7df-a2033618af75","Type":"ContainerStarted","Data":"a7b4ec7b882480a2f1fb6e9739094cc0ff7f677853b24c4c6ef6421ff70a27a8"} Feb 25 13:36:38 crc kubenswrapper[4815]: I0225 13:36:38.420216 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 25 13:36:38 crc kubenswrapper[4815]: I0225 13:36:38.423039 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-hktjt" event={"ID":"4f6d715d-8d19-48be-b617-7bf2ea9b6098","Type":"ContainerStarted","Data":"dceb592aa80af9ffa23c7df5f0588a5649cccfea28db0235b843f8e1cdec1dc3"} Feb 25 13:36:38 crc kubenswrapper[4815]: I0225 13:36:38.423702 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:38 crc kubenswrapper[4815]: I0225 13:36:38.426930 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" event={"ID":"5926509f-0818-4ec5-96d3-ee2e9cb1504f","Type":"ContainerStarted","Data":"654c23f91155839473b29d5ba5fbf77373d8ee16b6c98834fa987951017e9173"} Feb 25 13:36:38 crc kubenswrapper[4815]: I0225 13:36:38.427632 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:38 crc kubenswrapper[4815]: I0225 13:36:38.453930 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.271748431 podStartE2EDuration="3.453908868s" podCreationTimestamp="2026-02-25 13:36:35 +0000 UTC" firstStartedPulling="2026-02-25 13:36:36.556298209 +0000 UTC m=+954.357396263" lastFinishedPulling="2026-02-25 13:36:37.738458646 +0000 UTC m=+955.539556700" observedRunningTime="2026-02-25 13:36:38.444329402 +0000 UTC m=+956.245427476" watchObservedRunningTime="2026-02-25 13:36:38.453908868 +0000 UTC m=+956.255006932" Feb 25 13:36:38 crc kubenswrapper[4815]: I0225 13:36:38.465951 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" podStartSLOduration=3.726157696 podStartE2EDuration="4.46592753s" podCreationTimestamp="2026-02-25 13:36:34 +0000 UTC" firstStartedPulling="2026-02-25 13:36:35.342389751 +0000 UTC m=+953.143487805" lastFinishedPulling="2026-02-25 13:36:36.082159585 +0000 UTC m=+953.883257639" observedRunningTime="2026-02-25 13:36:38.462422461 +0000 UTC m=+956.263520515" watchObservedRunningTime="2026-02-25 13:36:38.46592753 +0000 UTC m=+956.267025584" Feb 25 13:36:38 crc kubenswrapper[4815]: I0225 13:36:38.486627 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-hktjt" podStartSLOduration=4.072084087 podStartE2EDuration="4.486601779s" podCreationTimestamp="2026-02-25 13:36:34 +0000 UTC" firstStartedPulling="2026-02-25 13:36:35.948739991 +0000 UTC m=+953.749838045" lastFinishedPulling="2026-02-25 13:36:36.363257683 +0000 UTC m=+954.164355737" observedRunningTime="2026-02-25 13:36:38.48274788 +0000 UTC m=+956.283845944" watchObservedRunningTime="2026-02-25 13:36:38.486601779 +0000 UTC m=+956.287699873" Feb 25 13:36:38 crc kubenswrapper[4815]: I0225 13:36:38.560454 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 25 13:36:38 crc kubenswrapper[4815]: I0225 13:36:38.560905 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 25 13:36:40 crc kubenswrapper[4815]: I0225 13:36:40.073425 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:40 crc kubenswrapper[4815]: I0225 13:36:40.073827 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:41 crc kubenswrapper[4815]: I0225 13:36:41.388176 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 25 13:36:41 crc kubenswrapper[4815]: I0225 13:36:41.489420 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="52d1a204-9bb6-40e1-a63c-09ffe6016cd9" containerName="galera" probeResult="failure" output=< Feb 25 13:36:41 crc kubenswrapper[4815]: wsrep_local_state_comment (Joined) differs from Synced Feb 25 13:36:41 crc kubenswrapper[4815]: > Feb 25 13:36:42 crc kubenswrapper[4815]: I0225 13:36:42.485246 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:42 crc kubenswrapper[4815]: I0225 13:36:42.581053 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="c96edb5d-f9a9-4dc1-a9c2-e3def3efb850" containerName="galera" probeResult="failure" output=< Feb 25 13:36:42 crc kubenswrapper[4815]: wsrep_local_state_comment (Joined) differs from Synced Feb 25 13:36:42 crc kubenswrapper[4815]: > Feb 25 13:36:42 crc kubenswrapper[4815]: I0225 13:36:42.829446 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 25 13:36:42 crc kubenswrapper[4815]: I0225 13:36:42.895886 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-74nkx"] Feb 25 13:36:42 crc kubenswrapper[4815]: I0225 13:36:42.896084 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" podUID="5926509f-0818-4ec5-96d3-ee2e9cb1504f" containerName="dnsmasq-dns" containerID="cri-o://654c23f91155839473b29d5ba5fbf77373d8ee16b6c98834fa987951017e9173" gracePeriod=10 Feb 25 13:36:42 crc kubenswrapper[4815]: I0225 13:36:42.898170 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:42 crc kubenswrapper[4815]: I0225 13:36:42.951494 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-kj6n4"] Feb 25 13:36:42 crc kubenswrapper[4815]: I0225 13:36:42.952835 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:42 crc kubenswrapper[4815]: I0225 13:36:42.983456 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-kj6n4"] Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.039993 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2824z\" (UniqueName: \"kubernetes.io/projected/09095c00-a8b0-4c47-84e5-4db2eb1d4309-kube-api-access-2824z\") pod \"dnsmasq-dns-b8fbc5445-kj6n4\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.040048 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-kj6n4\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.040078 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-kj6n4\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.040105 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-kj6n4\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.040141 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-config\") pod \"dnsmasq-dns-b8fbc5445-kj6n4\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.149276 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2824z\" (UniqueName: \"kubernetes.io/projected/09095c00-a8b0-4c47-84e5-4db2eb1d4309-kube-api-access-2824z\") pod \"dnsmasq-dns-b8fbc5445-kj6n4\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.149341 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-kj6n4\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.149371 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-kj6n4\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.149400 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-kj6n4\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.149442 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-config\") pod \"dnsmasq-dns-b8fbc5445-kj6n4\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.150351 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-config\") pod \"dnsmasq-dns-b8fbc5445-kj6n4\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.151237 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-kj6n4\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.152090 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-kj6n4\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.156808 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-kj6n4\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.185656 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2824z\" (UniqueName: \"kubernetes.io/projected/09095c00-a8b0-4c47-84e5-4db2eb1d4309-kube-api-access-2824z\") pod \"dnsmasq-dns-b8fbc5445-kj6n4\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.275921 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.462884 4815 generic.go:334] "Generic (PLEG): container finished" podID="5926509f-0818-4ec5-96d3-ee2e9cb1504f" containerID="654c23f91155839473b29d5ba5fbf77373d8ee16b6c98834fa987951017e9173" exitCode=0 Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.463180 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" event={"ID":"5926509f-0818-4ec5-96d3-ee2e9cb1504f","Type":"ContainerDied","Data":"654c23f91155839473b29d5ba5fbf77373d8ee16b6c98834fa987951017e9173"} Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.670329 4815 scope.go:117] "RemoveContainer" containerID="43314d9c57872caad1ed75670b9302c7293fc3dd8c04ba73a133bfa03c431ed3" Feb 25 13:36:43 crc kubenswrapper[4815]: I0225 13:36:43.705481 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-kj6n4"] Feb 25 13:36:43 crc kubenswrapper[4815]: W0225 13:36:43.750718 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09095c00_a8b0_4c47_84e5_4db2eb1d4309.slice/crio-41ba47c4105756c83948b222e4d7bbfd1ce535823a8c281931c3c84cbf3bd4a1 WatchSource:0}: Error finding container 41ba47c4105756c83948b222e4d7bbfd1ce535823a8c281931c3c84cbf3bd4a1: Status 404 returned error can't find the container with id 41ba47c4105756c83948b222e4d7bbfd1ce535823a8c281931c3c84cbf3bd4a1 Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.039617 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.050790 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.056882 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.056930 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.057053 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-pxcjf" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.057443 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.066389 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.167780 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5z6r\" (UniqueName: \"kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-kube-api-access-z5z6r\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.168046 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3941650f-c99b-4043-9137-e4c43361f93f-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.168165 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3941650f-c99b-4043-9137-e4c43361f93f-lock\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.168240 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.168324 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3941650f-c99b-4043-9137-e4c43361f93f-cache\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.168447 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.269577 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5z6r\" (UniqueName: \"kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-kube-api-access-z5z6r\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.269938 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3941650f-c99b-4043-9137-e4c43361f93f-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.270129 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3941650f-c99b-4043-9137-e4c43361f93f-lock\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.270273 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.270395 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3941650f-c99b-4043-9137-e4c43361f93f-cache\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.270554 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.270665 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: E0225 13:36:44.270700 4815 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 25 13:36:44 crc kubenswrapper[4815]: E0225 13:36:44.270718 4815 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.270788 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3941650f-c99b-4043-9137-e4c43361f93f-lock\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.270943 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3941650f-c99b-4043-9137-e4c43361f93f-cache\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: E0225 13:36:44.270998 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift podName:3941650f-c99b-4043-9137-e4c43361f93f nodeName:}" failed. No retries permitted until 2026-02-25 13:36:44.770973536 +0000 UTC m=+962.572071600 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift") pod "swift-storage-0" (UID: "3941650f-c99b-4043-9137-e4c43361f93f") : configmap "swift-ring-files" not found Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.276490 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3941650f-c99b-4043-9137-e4c43361f93f-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.295103 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5z6r\" (UniqueName: \"kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-kube-api-access-z5z6r\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.296532 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.473472 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" event={"ID":"09095c00-a8b0-4c47-84e5-4db2eb1d4309","Type":"ContainerStarted","Data":"7422c646f8f65fd0c5a074ce28df635475cd197cb888b59eeaf7567c25aac492"} Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.473758 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" event={"ID":"09095c00-a8b0-4c47-84e5-4db2eb1d4309","Type":"ContainerStarted","Data":"41ba47c4105756c83948b222e4d7bbfd1ce535823a8c281931c3c84cbf3bd4a1"} Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.525897 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-zplf2"] Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.527046 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.529837 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.530020 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.530335 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.544708 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-zplf2"] Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.576977 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-swiftconf\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.577041 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-combined-ca-bundle\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.577088 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-scripts\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.577136 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-etc-swift\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.577239 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-dispersionconf\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.577274 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fvcr\" (UniqueName: \"kubernetes.io/projected/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-kube-api-access-9fvcr\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.577317 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-ring-data-devices\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.679689 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-dispersionconf\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.680940 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fvcr\" (UniqueName: \"kubernetes.io/projected/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-kube-api-access-9fvcr\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.681139 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-ring-data-devices\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.681276 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-swiftconf\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.681405 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-combined-ca-bundle\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.681573 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-scripts\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.681759 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-etc-swift\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.682129 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-ring-data-devices\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.682426 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-etc-swift\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.683123 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-scripts\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.688623 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-dispersionconf\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.690091 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-swiftconf\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.690304 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-combined-ca-bundle\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.700226 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fvcr\" (UniqueName: \"kubernetes.io/projected/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-kube-api-access-9fvcr\") pod \"swift-ring-rebalance-zplf2\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.758985 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.783020 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:44 crc kubenswrapper[4815]: E0225 13:36:44.783202 4815 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 25 13:36:44 crc kubenswrapper[4815]: E0225 13:36:44.783223 4815 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 25 13:36:44 crc kubenswrapper[4815]: E0225 13:36:44.783275 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift podName:3941650f-c99b-4043-9137-e4c43361f93f nodeName:}" failed. No retries permitted until 2026-02-25 13:36:45.783259479 +0000 UTC m=+963.584357533 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift") pod "swift-storage-0" (UID: "3941650f-c99b-4043-9137-e4c43361f93f") : configmap "swift-ring-files" not found Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.866358 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.883706 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-dns-svc\") pod \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\" (UID: \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\") " Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.883896 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-ovsdbserver-nb\") pod \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\" (UID: \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\") " Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.883951 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-config\") pod \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\" (UID: \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\") " Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.883974 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tknfg\" (UniqueName: \"kubernetes.io/projected/5926509f-0818-4ec5-96d3-ee2e9cb1504f-kube-api-access-tknfg\") pod \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\" (UID: \"5926509f-0818-4ec5-96d3-ee2e9cb1504f\") " Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.890449 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5926509f-0818-4ec5-96d3-ee2e9cb1504f-kube-api-access-tknfg" (OuterVolumeSpecName: "kube-api-access-tknfg") pod "5926509f-0818-4ec5-96d3-ee2e9cb1504f" (UID: "5926509f-0818-4ec5-96d3-ee2e9cb1504f"). InnerVolumeSpecName "kube-api-access-tknfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.933009 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5926509f-0818-4ec5-96d3-ee2e9cb1504f" (UID: "5926509f-0818-4ec5-96d3-ee2e9cb1504f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.944224 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5926509f-0818-4ec5-96d3-ee2e9cb1504f" (UID: "5926509f-0818-4ec5-96d3-ee2e9cb1504f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.963836 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-config" (OuterVolumeSpecName: "config") pod "5926509f-0818-4ec5-96d3-ee2e9cb1504f" (UID: "5926509f-0818-4ec5-96d3-ee2e9cb1504f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.986470 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.986542 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.986556 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5926509f-0818-4ec5-96d3-ee2e9cb1504f-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:44 crc kubenswrapper[4815]: I0225 13:36:44.986567 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tknfg\" (UniqueName: \"kubernetes.io/projected/5926509f-0818-4ec5-96d3-ee2e9cb1504f-kube-api-access-tknfg\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:45 crc kubenswrapper[4815]: I0225 13:36:45.333716 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:45 crc kubenswrapper[4815]: I0225 13:36:45.404434 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-zplf2"] Feb 25 13:36:45 crc kubenswrapper[4815]: I0225 13:36:45.481662 4815 generic.go:334] "Generic (PLEG): container finished" podID="09095c00-a8b0-4c47-84e5-4db2eb1d4309" containerID="7422c646f8f65fd0c5a074ce28df635475cd197cb888b59eeaf7567c25aac492" exitCode=0 Feb 25 13:36:45 crc kubenswrapper[4815]: I0225 13:36:45.481794 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" event={"ID":"09095c00-a8b0-4c47-84e5-4db2eb1d4309","Type":"ContainerDied","Data":"7422c646f8f65fd0c5a074ce28df635475cd197cb888b59eeaf7567c25aac492"} Feb 25 13:36:45 crc kubenswrapper[4815]: I0225 13:36:45.482987 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zplf2" event={"ID":"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7","Type":"ContainerStarted","Data":"860d930d6766f6f79181fd8d9396f3b1385e634c4154223530dfcf1d9a41cfb7"} Feb 25 13:36:45 crc kubenswrapper[4815]: I0225 13:36:45.489729 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" event={"ID":"5926509f-0818-4ec5-96d3-ee2e9cb1504f","Type":"ContainerDied","Data":"fd3ebc7986cc63fc3ed05f1fc7ce820433f5b0dd15e0780f2138d24c2c5ecc07"} Feb 25 13:36:45 crc kubenswrapper[4815]: I0225 13:36:45.489769 4815 scope.go:117] "RemoveContainer" containerID="654c23f91155839473b29d5ba5fbf77373d8ee16b6c98834fa987951017e9173" Feb 25 13:36:45 crc kubenswrapper[4815]: I0225 13:36:45.489804 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-74nkx" Feb 25 13:36:45 crc kubenswrapper[4815]: I0225 13:36:45.594212 4815 scope.go:117] "RemoveContainer" containerID="8c3dd5b64e6d21e8a1ef6438d5038832d8c2b167a70c5da5403d26cdcc959566" Feb 25 13:36:45 crc kubenswrapper[4815]: I0225 13:36:45.642246 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-74nkx"] Feb 25 13:36:45 crc kubenswrapper[4815]: I0225 13:36:45.648675 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-74nkx"] Feb 25 13:36:45 crc kubenswrapper[4815]: I0225 13:36:45.804829 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:45 crc kubenswrapper[4815]: E0225 13:36:45.805033 4815 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 25 13:36:45 crc kubenswrapper[4815]: E0225 13:36:45.805060 4815 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 25 13:36:45 crc kubenswrapper[4815]: E0225 13:36:45.805118 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift podName:3941650f-c99b-4043-9137-e4c43361f93f nodeName:}" failed. No retries permitted until 2026-02-25 13:36:47.805097832 +0000 UTC m=+965.606195886 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift") pod "swift-storage-0" (UID: "3941650f-c99b-4043-9137-e4c43361f93f") : configmap "swift-ring-files" not found Feb 25 13:36:46 crc kubenswrapper[4815]: I0225 13:36:46.501573 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" event={"ID":"09095c00-a8b0-4c47-84e5-4db2eb1d4309","Type":"ContainerStarted","Data":"cd5bddfd310cfee75c571904678ee904d02b05cac2419967fc5d5a8ca4c24d27"} Feb 25 13:36:46 crc kubenswrapper[4815]: I0225 13:36:46.501961 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:46 crc kubenswrapper[4815]: I0225 13:36:46.951902 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5926509f-0818-4ec5-96d3-ee2e9cb1504f" path="/var/lib/kubelet/pods/5926509f-0818-4ec5-96d3-ee2e9cb1504f/volumes" Feb 25 13:36:47 crc kubenswrapper[4815]: I0225 13:36:47.838748 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:47 crc kubenswrapper[4815]: E0225 13:36:47.838963 4815 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 25 13:36:47 crc kubenswrapper[4815]: E0225 13:36:47.838990 4815 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 25 13:36:47 crc kubenswrapper[4815]: E0225 13:36:47.839066 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift podName:3941650f-c99b-4043-9137-e4c43361f93f nodeName:}" failed. No retries permitted until 2026-02-25 13:36:51.839040615 +0000 UTC m=+969.640138709 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift") pod "swift-storage-0" (UID: "3941650f-c99b-4043-9137-e4c43361f93f") : configmap "swift-ring-files" not found Feb 25 13:36:48 crc kubenswrapper[4815]: I0225 13:36:48.642276 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 25 13:36:48 crc kubenswrapper[4815]: I0225 13:36:48.674529 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" podStartSLOduration=6.6744928869999995 podStartE2EDuration="6.674492887s" podCreationTimestamp="2026-02-25 13:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:36:46.522107372 +0000 UTC m=+964.323205426" watchObservedRunningTime="2026-02-25 13:36:48.674492887 +0000 UTC m=+966.475590941" Feb 25 13:36:49 crc kubenswrapper[4815]: I0225 13:36:49.530407 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zplf2" event={"ID":"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7","Type":"ContainerStarted","Data":"0fdd4260540bf1feb5f5070423d97484add03687a692d07a94d34439324625a7"} Feb 25 13:36:49 crc kubenswrapper[4815]: I0225 13:36:49.550990 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-zplf2" podStartSLOduration=2.3221405219999998 podStartE2EDuration="5.550965615s" podCreationTimestamp="2026-02-25 13:36:44 +0000 UTC" firstStartedPulling="2026-02-25 13:36:45.375868286 +0000 UTC m=+963.176966380" lastFinishedPulling="2026-02-25 13:36:48.604693419 +0000 UTC m=+966.405791473" observedRunningTime="2026-02-25 13:36:49.549927104 +0000 UTC m=+967.351025188" watchObservedRunningTime="2026-02-25 13:36:49.550965615 +0000 UTC m=+967.352063709" Feb 25 13:36:50 crc kubenswrapper[4815]: I0225 13:36:50.188170 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.341768 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.343003 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.512096 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-e541-account-create-update-7769t"] Feb 25 13:36:51 crc kubenswrapper[4815]: E0225 13:36:51.512673 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5926509f-0818-4ec5-96d3-ee2e9cb1504f" containerName="init" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.512697 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="5926509f-0818-4ec5-96d3-ee2e9cb1504f" containerName="init" Feb 25 13:36:51 crc kubenswrapper[4815]: E0225 13:36:51.512720 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5926509f-0818-4ec5-96d3-ee2e9cb1504f" containerName="dnsmasq-dns" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.512749 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="5926509f-0818-4ec5-96d3-ee2e9cb1504f" containerName="dnsmasq-dns" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.513040 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="5926509f-0818-4ec5-96d3-ee2e9cb1504f" containerName="dnsmasq-dns" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.513878 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e541-account-create-update-7769t" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.517152 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.521980 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e541-account-create-update-7769t"] Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.578292 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-gp6dc"] Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.579359 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gp6dc" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.584484 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gp6dc"] Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.608485 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85f6g\" (UniqueName: \"kubernetes.io/projected/a24929f6-66ff-415c-b539-a9e6ac507f1d-kube-api-access-85f6g\") pod \"keystone-db-create-gp6dc\" (UID: \"a24929f6-66ff-415c-b539-a9e6ac507f1d\") " pod="openstack/keystone-db-create-gp6dc" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.608564 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a24929f6-66ff-415c-b539-a9e6ac507f1d-operator-scripts\") pod \"keystone-db-create-gp6dc\" (UID: \"a24929f6-66ff-415c-b539-a9e6ac507f1d\") " pod="openstack/keystone-db-create-gp6dc" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.608688 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84c1760e-09f1-4f63-ae74-8c8ba4af49be-operator-scripts\") pod \"keystone-e541-account-create-update-7769t\" (UID: \"84c1760e-09f1-4f63-ae74-8c8ba4af49be\") " pod="openstack/keystone-e541-account-create-update-7769t" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.608807 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cghk\" (UniqueName: \"kubernetes.io/projected/84c1760e-09f1-4f63-ae74-8c8ba4af49be-kube-api-access-8cghk\") pod \"keystone-e541-account-create-update-7769t\" (UID: \"84c1760e-09f1-4f63-ae74-8c8ba4af49be\") " pod="openstack/keystone-e541-account-create-update-7769t" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.669051 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-4xctv"] Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.670261 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-4xctv" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.675814 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-4xctv"] Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.710896 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85f6g\" (UniqueName: \"kubernetes.io/projected/a24929f6-66ff-415c-b539-a9e6ac507f1d-kube-api-access-85f6g\") pod \"keystone-db-create-gp6dc\" (UID: \"a24929f6-66ff-415c-b539-a9e6ac507f1d\") " pod="openstack/keystone-db-create-gp6dc" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.710976 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a24929f6-66ff-415c-b539-a9e6ac507f1d-operator-scripts\") pod \"keystone-db-create-gp6dc\" (UID: \"a24929f6-66ff-415c-b539-a9e6ac507f1d\") " pod="openstack/keystone-db-create-gp6dc" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.711013 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84c1760e-09f1-4f63-ae74-8c8ba4af49be-operator-scripts\") pod \"keystone-e541-account-create-update-7769t\" (UID: \"84c1760e-09f1-4f63-ae74-8c8ba4af49be\") " pod="openstack/keystone-e541-account-create-update-7769t" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.711038 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50e8c23e-f13a-4fb9-b223-9330b3d1d36b-operator-scripts\") pod \"placement-db-create-4xctv\" (UID: \"50e8c23e-f13a-4fb9-b223-9330b3d1d36b\") " pod="openstack/placement-db-create-4xctv" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.711080 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cghk\" (UniqueName: \"kubernetes.io/projected/84c1760e-09f1-4f63-ae74-8c8ba4af49be-kube-api-access-8cghk\") pod \"keystone-e541-account-create-update-7769t\" (UID: \"84c1760e-09f1-4f63-ae74-8c8ba4af49be\") " pod="openstack/keystone-e541-account-create-update-7769t" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.711128 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzmws\" (UniqueName: \"kubernetes.io/projected/50e8c23e-f13a-4fb9-b223-9330b3d1d36b-kube-api-access-vzmws\") pod \"placement-db-create-4xctv\" (UID: \"50e8c23e-f13a-4fb9-b223-9330b3d1d36b\") " pod="openstack/placement-db-create-4xctv" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.711730 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a24929f6-66ff-415c-b539-a9e6ac507f1d-operator-scripts\") pod \"keystone-db-create-gp6dc\" (UID: \"a24929f6-66ff-415c-b539-a9e6ac507f1d\") " pod="openstack/keystone-db-create-gp6dc" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.711894 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84c1760e-09f1-4f63-ae74-8c8ba4af49be-operator-scripts\") pod \"keystone-e541-account-create-update-7769t\" (UID: \"84c1760e-09f1-4f63-ae74-8c8ba4af49be\") " pod="openstack/keystone-e541-account-create-update-7769t" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.728962 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85f6g\" (UniqueName: \"kubernetes.io/projected/a24929f6-66ff-415c-b539-a9e6ac507f1d-kube-api-access-85f6g\") pod \"keystone-db-create-gp6dc\" (UID: \"a24929f6-66ff-415c-b539-a9e6ac507f1d\") " pod="openstack/keystone-db-create-gp6dc" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.735400 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cghk\" (UniqueName: \"kubernetes.io/projected/84c1760e-09f1-4f63-ae74-8c8ba4af49be-kube-api-access-8cghk\") pod \"keystone-e541-account-create-update-7769t\" (UID: \"84c1760e-09f1-4f63-ae74-8c8ba4af49be\") " pod="openstack/keystone-e541-account-create-update-7769t" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.812870 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50e8c23e-f13a-4fb9-b223-9330b3d1d36b-operator-scripts\") pod \"placement-db-create-4xctv\" (UID: \"50e8c23e-f13a-4fb9-b223-9330b3d1d36b\") " pod="openstack/placement-db-create-4xctv" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.812967 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzmws\" (UniqueName: \"kubernetes.io/projected/50e8c23e-f13a-4fb9-b223-9330b3d1d36b-kube-api-access-vzmws\") pod \"placement-db-create-4xctv\" (UID: \"50e8c23e-f13a-4fb9-b223-9330b3d1d36b\") " pod="openstack/placement-db-create-4xctv" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.814488 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50e8c23e-f13a-4fb9-b223-9330b3d1d36b-operator-scripts\") pod \"placement-db-create-4xctv\" (UID: \"50e8c23e-f13a-4fb9-b223-9330b3d1d36b\") " pod="openstack/placement-db-create-4xctv" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.814716 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-29cc-account-create-update-rf4f5"] Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.815915 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-29cc-account-create-update-rf4f5" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.817856 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.830216 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-29cc-account-create-update-rf4f5"] Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.836467 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzmws\" (UniqueName: \"kubernetes.io/projected/50e8c23e-f13a-4fb9-b223-9330b3d1d36b-kube-api-access-vzmws\") pod \"placement-db-create-4xctv\" (UID: \"50e8c23e-f13a-4fb9-b223-9330b3d1d36b\") " pod="openstack/placement-db-create-4xctv" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.857919 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e541-account-create-update-7769t" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.902861 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gp6dc" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.926335 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc782\" (UniqueName: \"kubernetes.io/projected/64ed35f4-2af0-4d3e-84a7-6160ac5b9185-kube-api-access-pc782\") pod \"placement-29cc-account-create-update-rf4f5\" (UID: \"64ed35f4-2af0-4d3e-84a7-6160ac5b9185\") " pod="openstack/placement-29cc-account-create-update-rf4f5" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.926536 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.926566 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64ed35f4-2af0-4d3e-84a7-6160ac5b9185-operator-scripts\") pod \"placement-29cc-account-create-update-rf4f5\" (UID: \"64ed35f4-2af0-4d3e-84a7-6160ac5b9185\") " pod="openstack/placement-29cc-account-create-update-rf4f5" Feb 25 13:36:51 crc kubenswrapper[4815]: E0225 13:36:51.926752 4815 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 25 13:36:51 crc kubenswrapper[4815]: E0225 13:36:51.926766 4815 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 25 13:36:51 crc kubenswrapper[4815]: E0225 13:36:51.926807 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift podName:3941650f-c99b-4043-9137-e4c43361f93f nodeName:}" failed. No retries permitted until 2026-02-25 13:36:59.926791045 +0000 UTC m=+977.727889099 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift") pod "swift-storage-0" (UID: "3941650f-c99b-4043-9137-e4c43361f93f") : configmap "swift-ring-files" not found Feb 25 13:36:51 crc kubenswrapper[4815]: I0225 13:36:51.984193 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-4xctv" Feb 25 13:36:52 crc kubenswrapper[4815]: I0225 13:36:52.028643 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc782\" (UniqueName: \"kubernetes.io/projected/64ed35f4-2af0-4d3e-84a7-6160ac5b9185-kube-api-access-pc782\") pod \"placement-29cc-account-create-update-rf4f5\" (UID: \"64ed35f4-2af0-4d3e-84a7-6160ac5b9185\") " pod="openstack/placement-29cc-account-create-update-rf4f5" Feb 25 13:36:52 crc kubenswrapper[4815]: I0225 13:36:52.028823 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64ed35f4-2af0-4d3e-84a7-6160ac5b9185-operator-scripts\") pod \"placement-29cc-account-create-update-rf4f5\" (UID: \"64ed35f4-2af0-4d3e-84a7-6160ac5b9185\") " pod="openstack/placement-29cc-account-create-update-rf4f5" Feb 25 13:36:52 crc kubenswrapper[4815]: I0225 13:36:52.029463 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64ed35f4-2af0-4d3e-84a7-6160ac5b9185-operator-scripts\") pod \"placement-29cc-account-create-update-rf4f5\" (UID: \"64ed35f4-2af0-4d3e-84a7-6160ac5b9185\") " pod="openstack/placement-29cc-account-create-update-rf4f5" Feb 25 13:36:52 crc kubenswrapper[4815]: I0225 13:36:52.057406 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc782\" (UniqueName: \"kubernetes.io/projected/64ed35f4-2af0-4d3e-84a7-6160ac5b9185-kube-api-access-pc782\") pod \"placement-29cc-account-create-update-rf4f5\" (UID: \"64ed35f4-2af0-4d3e-84a7-6160ac5b9185\") " pod="openstack/placement-29cc-account-create-update-rf4f5" Feb 25 13:36:52 crc kubenswrapper[4815]: I0225 13:36:52.130763 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-29cc-account-create-update-rf4f5" Feb 25 13:36:52 crc kubenswrapper[4815]: W0225 13:36:52.428111 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84c1760e_09f1_4f63_ae74_8c8ba4af49be.slice/crio-ae55c307c6d91af918bf42b4405a4c32a40efb4d0094a9218c94cd9d7139388c WatchSource:0}: Error finding container ae55c307c6d91af918bf42b4405a4c32a40efb4d0094a9218c94cd9d7139388c: Status 404 returned error can't find the container with id ae55c307c6d91af918bf42b4405a4c32a40efb4d0094a9218c94cd9d7139388c Feb 25 13:36:52 crc kubenswrapper[4815]: I0225 13:36:52.433168 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e541-account-create-update-7769t"] Feb 25 13:36:52 crc kubenswrapper[4815]: I0225 13:36:52.481047 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-4xctv"] Feb 25 13:36:52 crc kubenswrapper[4815]: W0225 13:36:52.492253 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50e8c23e_f13a_4fb9_b223_9330b3d1d36b.slice/crio-018f5d5f03ccd378da703951e73bb703ae373c1b137d005da247565cbfbbee5b WatchSource:0}: Error finding container 018f5d5f03ccd378da703951e73bb703ae373c1b137d005da247565cbfbbee5b: Status 404 returned error can't find the container with id 018f5d5f03ccd378da703951e73bb703ae373c1b137d005da247565cbfbbee5b Feb 25 13:36:52 crc kubenswrapper[4815]: I0225 13:36:52.493177 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gp6dc"] Feb 25 13:36:52 crc kubenswrapper[4815]: W0225 13:36:52.496064 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda24929f6_66ff_415c_b539_a9e6ac507f1d.slice/crio-ffe481017dd6db707c56b4e05c0ebffff5f260e6c7ae470379fbfc924033c4ef WatchSource:0}: Error finding container ffe481017dd6db707c56b4e05c0ebffff5f260e6c7ae470379fbfc924033c4ef: Status 404 returned error can't find the container with id ffe481017dd6db707c56b4e05c0ebffff5f260e6c7ae470379fbfc924033c4ef Feb 25 13:36:52 crc kubenswrapper[4815]: I0225 13:36:52.555939 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gp6dc" event={"ID":"a24929f6-66ff-415c-b539-a9e6ac507f1d","Type":"ContainerStarted","Data":"ffe481017dd6db707c56b4e05c0ebffff5f260e6c7ae470379fbfc924033c4ef"} Feb 25 13:36:52 crc kubenswrapper[4815]: I0225 13:36:52.557157 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e541-account-create-update-7769t" event={"ID":"84c1760e-09f1-4f63-ae74-8c8ba4af49be","Type":"ContainerStarted","Data":"ae55c307c6d91af918bf42b4405a4c32a40efb4d0094a9218c94cd9d7139388c"} Feb 25 13:36:52 crc kubenswrapper[4815]: I0225 13:36:52.558598 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-4xctv" event={"ID":"50e8c23e-f13a-4fb9-b223-9330b3d1d36b","Type":"ContainerStarted","Data":"018f5d5f03ccd378da703951e73bb703ae373c1b137d005da247565cbfbbee5b"} Feb 25 13:36:52 crc kubenswrapper[4815]: I0225 13:36:52.626382 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-29cc-account-create-update-rf4f5"] Feb 25 13:36:52 crc kubenswrapper[4815]: W0225 13:36:52.632754 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64ed35f4_2af0_4d3e_84a7_6160ac5b9185.slice/crio-2c26ffd234dcc02874056d518d84acea0a897401521108b6c5ea888b6c1654b4 WatchSource:0}: Error finding container 2c26ffd234dcc02874056d518d84acea0a897401521108b6c5ea888b6c1654b4: Status 404 returned error can't find the container with id 2c26ffd234dcc02874056d518d84acea0a897401521108b6c5ea888b6c1654b4 Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.277700 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.338819 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-hktjt"] Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.339084 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-hktjt" podUID="4f6d715d-8d19-48be-b617-7bf2ea9b6098" containerName="dnsmasq-dns" containerID="cri-o://dceb592aa80af9ffa23c7df5f0588a5649cccfea28db0235b843f8e1cdec1dc3" gracePeriod=10 Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.569676 4815 generic.go:334] "Generic (PLEG): container finished" podID="64ed35f4-2af0-4d3e-84a7-6160ac5b9185" containerID="32bb1bba6e9f9bc6ce15dd3c3911076bc6a8ef51fc9788763a3d5e23b1ef529b" exitCode=0 Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.570361 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-29cc-account-create-update-rf4f5" event={"ID":"64ed35f4-2af0-4d3e-84a7-6160ac5b9185","Type":"ContainerDied","Data":"32bb1bba6e9f9bc6ce15dd3c3911076bc6a8ef51fc9788763a3d5e23b1ef529b"} Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.570395 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-29cc-account-create-update-rf4f5" event={"ID":"64ed35f4-2af0-4d3e-84a7-6160ac5b9185","Type":"ContainerStarted","Data":"2c26ffd234dcc02874056d518d84acea0a897401521108b6c5ea888b6c1654b4"} Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.571905 4815 generic.go:334] "Generic (PLEG): container finished" podID="50e8c23e-f13a-4fb9-b223-9330b3d1d36b" containerID="01e045a2656f40d52c581376e503f61bb812e34ebe0a3607a28c08beaf4adedf" exitCode=0 Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.572007 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-4xctv" event={"ID":"50e8c23e-f13a-4fb9-b223-9330b3d1d36b","Type":"ContainerDied","Data":"01e045a2656f40d52c581376e503f61bb812e34ebe0a3607a28c08beaf4adedf"} Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.574078 4815 generic.go:334] "Generic (PLEG): container finished" podID="4f6d715d-8d19-48be-b617-7bf2ea9b6098" containerID="dceb592aa80af9ffa23c7df5f0588a5649cccfea28db0235b843f8e1cdec1dc3" exitCode=0 Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.574116 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-hktjt" event={"ID":"4f6d715d-8d19-48be-b617-7bf2ea9b6098","Type":"ContainerDied","Data":"dceb592aa80af9ffa23c7df5f0588a5649cccfea28db0235b843f8e1cdec1dc3"} Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.575283 4815 generic.go:334] "Generic (PLEG): container finished" podID="a24929f6-66ff-415c-b539-a9e6ac507f1d" containerID="a5c9d925676cb371db86aa6cc24221d97964582b03c81997911af246f539ca24" exitCode=0 Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.575321 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gp6dc" event={"ID":"a24929f6-66ff-415c-b539-a9e6ac507f1d","Type":"ContainerDied","Data":"a5c9d925676cb371db86aa6cc24221d97964582b03c81997911af246f539ca24"} Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.577495 4815 generic.go:334] "Generic (PLEG): container finished" podID="84c1760e-09f1-4f63-ae74-8c8ba4af49be" containerID="920719f20c53bbaee0c90eb85e6b003ac665d58ab7ced9d807566981c86c04e7" exitCode=0 Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.577570 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e541-account-create-update-7769t" event={"ID":"84c1760e-09f1-4f63-ae74-8c8ba4af49be","Type":"ContainerDied","Data":"920719f20c53bbaee0c90eb85e6b003ac665d58ab7ced9d807566981c86c04e7"} Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.827371 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.965663 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-ovsdbserver-nb\") pod \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.965747 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-config\") pod \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.965872 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-ovsdbserver-sb\") pod \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.965895 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-dns-svc\") pod \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.966003 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2846\" (UniqueName: \"kubernetes.io/projected/4f6d715d-8d19-48be-b617-7bf2ea9b6098-kube-api-access-w2846\") pod \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\" (UID: \"4f6d715d-8d19-48be-b617-7bf2ea9b6098\") " Feb 25 13:36:53 crc kubenswrapper[4815]: I0225 13:36:53.985788 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f6d715d-8d19-48be-b617-7bf2ea9b6098-kube-api-access-w2846" (OuterVolumeSpecName: "kube-api-access-w2846") pod "4f6d715d-8d19-48be-b617-7bf2ea9b6098" (UID: "4f6d715d-8d19-48be-b617-7bf2ea9b6098"). InnerVolumeSpecName "kube-api-access-w2846". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:36:54 crc kubenswrapper[4815]: I0225 13:36:54.011810 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-config" (OuterVolumeSpecName: "config") pod "4f6d715d-8d19-48be-b617-7bf2ea9b6098" (UID: "4f6d715d-8d19-48be-b617-7bf2ea9b6098"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:54 crc kubenswrapper[4815]: I0225 13:36:54.018188 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4f6d715d-8d19-48be-b617-7bf2ea9b6098" (UID: "4f6d715d-8d19-48be-b617-7bf2ea9b6098"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:54 crc kubenswrapper[4815]: I0225 13:36:54.022769 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4f6d715d-8d19-48be-b617-7bf2ea9b6098" (UID: "4f6d715d-8d19-48be-b617-7bf2ea9b6098"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:54 crc kubenswrapper[4815]: I0225 13:36:54.023275 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4f6d715d-8d19-48be-b617-7bf2ea9b6098" (UID: "4f6d715d-8d19-48be-b617-7bf2ea9b6098"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:54 crc kubenswrapper[4815]: I0225 13:36:54.068132 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:54 crc kubenswrapper[4815]: I0225 13:36:54.068164 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:54 crc kubenswrapper[4815]: I0225 13:36:54.068182 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2846\" (UniqueName: \"kubernetes.io/projected/4f6d715d-8d19-48be-b617-7bf2ea9b6098-kube-api-access-w2846\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:54 crc kubenswrapper[4815]: I0225 13:36:54.068191 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:54 crc kubenswrapper[4815]: I0225 13:36:54.068202 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f6d715d-8d19-48be-b617-7bf2ea9b6098-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:54 crc kubenswrapper[4815]: I0225 13:36:54.588553 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-hktjt" event={"ID":"4f6d715d-8d19-48be-b617-7bf2ea9b6098","Type":"ContainerDied","Data":"cc6e372c493fe6ad5b38db578ed73ee78fdd863b0da5607ef9a6b9af08b53b20"} Feb 25 13:36:54 crc kubenswrapper[4815]: I0225 13:36:54.588606 4815 scope.go:117] "RemoveContainer" containerID="dceb592aa80af9ffa23c7df5f0588a5649cccfea28db0235b843f8e1cdec1dc3" Feb 25 13:36:54 crc kubenswrapper[4815]: I0225 13:36:54.588841 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-hktjt" Feb 25 13:36:54 crc kubenswrapper[4815]: I0225 13:36:54.641714 4815 scope.go:117] "RemoveContainer" containerID="5cfd11297a7ab195f05dd312402f3acc2d4e7fbbf06cdddd2e6c5330680e0176" Feb 25 13:36:54 crc kubenswrapper[4815]: I0225 13:36:54.642075 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-hktjt"] Feb 25 13:36:54 crc kubenswrapper[4815]: I0225 13:36:54.648539 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-hktjt"] Feb 25 13:36:54 crc kubenswrapper[4815]: I0225 13:36:54.948000 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f6d715d-8d19-48be-b617-7bf2ea9b6098" path="/var/lib/kubelet/pods/4f6d715d-8d19-48be-b617-7bf2ea9b6098/volumes" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.047282 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e541-account-create-update-7769t" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.179562 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-29cc-account-create-update-rf4f5" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.191793 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cghk\" (UniqueName: \"kubernetes.io/projected/84c1760e-09f1-4f63-ae74-8c8ba4af49be-kube-api-access-8cghk\") pod \"84c1760e-09f1-4f63-ae74-8c8ba4af49be\" (UID: \"84c1760e-09f1-4f63-ae74-8c8ba4af49be\") " Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.191870 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84c1760e-09f1-4f63-ae74-8c8ba4af49be-operator-scripts\") pod \"84c1760e-09f1-4f63-ae74-8c8ba4af49be\" (UID: \"84c1760e-09f1-4f63-ae74-8c8ba4af49be\") " Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.192496 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84c1760e-09f1-4f63-ae74-8c8ba4af49be-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "84c1760e-09f1-4f63-ae74-8c8ba4af49be" (UID: "84c1760e-09f1-4f63-ae74-8c8ba4af49be"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.192861 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gp6dc" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.199403 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84c1760e-09f1-4f63-ae74-8c8ba4af49be-kube-api-access-8cghk" (OuterVolumeSpecName: "kube-api-access-8cghk") pod "84c1760e-09f1-4f63-ae74-8c8ba4af49be" (UID: "84c1760e-09f1-4f63-ae74-8c8ba4af49be"). InnerVolumeSpecName "kube-api-access-8cghk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.203305 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-4xctv" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.293090 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pc782\" (UniqueName: \"kubernetes.io/projected/64ed35f4-2af0-4d3e-84a7-6160ac5b9185-kube-api-access-pc782\") pod \"64ed35f4-2af0-4d3e-84a7-6160ac5b9185\" (UID: \"64ed35f4-2af0-4d3e-84a7-6160ac5b9185\") " Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.293172 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64ed35f4-2af0-4d3e-84a7-6160ac5b9185-operator-scripts\") pod \"64ed35f4-2af0-4d3e-84a7-6160ac5b9185\" (UID: \"64ed35f4-2af0-4d3e-84a7-6160ac5b9185\") " Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.293313 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzmws\" (UniqueName: \"kubernetes.io/projected/50e8c23e-f13a-4fb9-b223-9330b3d1d36b-kube-api-access-vzmws\") pod \"50e8c23e-f13a-4fb9-b223-9330b3d1d36b\" (UID: \"50e8c23e-f13a-4fb9-b223-9330b3d1d36b\") " Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.293367 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a24929f6-66ff-415c-b539-a9e6ac507f1d-operator-scripts\") pod \"a24929f6-66ff-415c-b539-a9e6ac507f1d\" (UID: \"a24929f6-66ff-415c-b539-a9e6ac507f1d\") " Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.293388 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50e8c23e-f13a-4fb9-b223-9330b3d1d36b-operator-scripts\") pod \"50e8c23e-f13a-4fb9-b223-9330b3d1d36b\" (UID: \"50e8c23e-f13a-4fb9-b223-9330b3d1d36b\") " Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.293407 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85f6g\" (UniqueName: \"kubernetes.io/projected/a24929f6-66ff-415c-b539-a9e6ac507f1d-kube-api-access-85f6g\") pod \"a24929f6-66ff-415c-b539-a9e6ac507f1d\" (UID: \"a24929f6-66ff-415c-b539-a9e6ac507f1d\") " Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.293723 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cghk\" (UniqueName: \"kubernetes.io/projected/84c1760e-09f1-4f63-ae74-8c8ba4af49be-kube-api-access-8cghk\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.293742 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84c1760e-09f1-4f63-ae74-8c8ba4af49be-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.294215 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a24929f6-66ff-415c-b539-a9e6ac507f1d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a24929f6-66ff-415c-b539-a9e6ac507f1d" (UID: "a24929f6-66ff-415c-b539-a9e6ac507f1d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.294222 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50e8c23e-f13a-4fb9-b223-9330b3d1d36b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "50e8c23e-f13a-4fb9-b223-9330b3d1d36b" (UID: "50e8c23e-f13a-4fb9-b223-9330b3d1d36b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.294305 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64ed35f4-2af0-4d3e-84a7-6160ac5b9185-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "64ed35f4-2af0-4d3e-84a7-6160ac5b9185" (UID: "64ed35f4-2af0-4d3e-84a7-6160ac5b9185"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.297083 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a24929f6-66ff-415c-b539-a9e6ac507f1d-kube-api-access-85f6g" (OuterVolumeSpecName: "kube-api-access-85f6g") pod "a24929f6-66ff-415c-b539-a9e6ac507f1d" (UID: "a24929f6-66ff-415c-b539-a9e6ac507f1d"). InnerVolumeSpecName "kube-api-access-85f6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.297604 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50e8c23e-f13a-4fb9-b223-9330b3d1d36b-kube-api-access-vzmws" (OuterVolumeSpecName: "kube-api-access-vzmws") pod "50e8c23e-f13a-4fb9-b223-9330b3d1d36b" (UID: "50e8c23e-f13a-4fb9-b223-9330b3d1d36b"). InnerVolumeSpecName "kube-api-access-vzmws". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.297697 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64ed35f4-2af0-4d3e-84a7-6160ac5b9185-kube-api-access-pc782" (OuterVolumeSpecName: "kube-api-access-pc782") pod "64ed35f4-2af0-4d3e-84a7-6160ac5b9185" (UID: "64ed35f4-2af0-4d3e-84a7-6160ac5b9185"). InnerVolumeSpecName "kube-api-access-pc782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.396972 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64ed35f4-2af0-4d3e-84a7-6160ac5b9185-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.397080 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzmws\" (UniqueName: \"kubernetes.io/projected/50e8c23e-f13a-4fb9-b223-9330b3d1d36b-kube-api-access-vzmws\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.397113 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a24929f6-66ff-415c-b539-a9e6ac507f1d-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.397138 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/50e8c23e-f13a-4fb9-b223-9330b3d1d36b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.397165 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85f6g\" (UniqueName: \"kubernetes.io/projected/a24929f6-66ff-415c-b539-a9e6ac507f1d-kube-api-access-85f6g\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.397191 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pc782\" (UniqueName: \"kubernetes.io/projected/64ed35f4-2af0-4d3e-84a7-6160ac5b9185-kube-api-access-pc782\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.482976 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-st8v2"] Feb 25 13:36:55 crc kubenswrapper[4815]: E0225 13:36:55.483448 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50e8c23e-f13a-4fb9-b223-9330b3d1d36b" containerName="mariadb-database-create" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.483475 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="50e8c23e-f13a-4fb9-b223-9330b3d1d36b" containerName="mariadb-database-create" Feb 25 13:36:55 crc kubenswrapper[4815]: E0225 13:36:55.483540 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64ed35f4-2af0-4d3e-84a7-6160ac5b9185" containerName="mariadb-account-create-update" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.483554 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="64ed35f4-2af0-4d3e-84a7-6160ac5b9185" containerName="mariadb-account-create-update" Feb 25 13:36:55 crc kubenswrapper[4815]: E0225 13:36:55.483571 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f6d715d-8d19-48be-b617-7bf2ea9b6098" containerName="init" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.483580 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f6d715d-8d19-48be-b617-7bf2ea9b6098" containerName="init" Feb 25 13:36:55 crc kubenswrapper[4815]: E0225 13:36:55.483595 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84c1760e-09f1-4f63-ae74-8c8ba4af49be" containerName="mariadb-account-create-update" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.483606 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="84c1760e-09f1-4f63-ae74-8c8ba4af49be" containerName="mariadb-account-create-update" Feb 25 13:36:55 crc kubenswrapper[4815]: E0225 13:36:55.483624 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f6d715d-8d19-48be-b617-7bf2ea9b6098" containerName="dnsmasq-dns" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.483632 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f6d715d-8d19-48be-b617-7bf2ea9b6098" containerName="dnsmasq-dns" Feb 25 13:36:55 crc kubenswrapper[4815]: E0225 13:36:55.483647 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a24929f6-66ff-415c-b539-a9e6ac507f1d" containerName="mariadb-database-create" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.483656 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a24929f6-66ff-415c-b539-a9e6ac507f1d" containerName="mariadb-database-create" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.483866 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="84c1760e-09f1-4f63-ae74-8c8ba4af49be" containerName="mariadb-account-create-update" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.483901 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="a24929f6-66ff-415c-b539-a9e6ac507f1d" containerName="mariadb-database-create" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.483916 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="64ed35f4-2af0-4d3e-84a7-6160ac5b9185" containerName="mariadb-account-create-update" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.483929 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f6d715d-8d19-48be-b617-7bf2ea9b6098" containerName="dnsmasq-dns" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.483950 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="50e8c23e-f13a-4fb9-b223-9330b3d1d36b" containerName="mariadb-database-create" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.484590 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-st8v2" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.495460 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-st8v2"] Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.582804 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-df70-account-create-update-s4qnr"] Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.583891 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-df70-account-create-update-s4qnr" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.587156 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.594409 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-df70-account-create-update-s4qnr"] Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.599971 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80276532-5b43-42b3-86db-de225388a3db-operator-scripts\") pod \"glance-db-create-st8v2\" (UID: \"80276532-5b43-42b3-86db-de225388a3db\") " pod="openstack/glance-db-create-st8v2" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.600126 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfgzf\" (UniqueName: \"kubernetes.io/projected/80276532-5b43-42b3-86db-de225388a3db-kube-api-access-qfgzf\") pod \"glance-db-create-st8v2\" (UID: \"80276532-5b43-42b3-86db-de225388a3db\") " pod="openstack/glance-db-create-st8v2" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.612744 4815 generic.go:334] "Generic (PLEG): container finished" podID="0b98e7b2-3068-4f59-b25f-4755bd752b49" containerID="960d7fcb921daf08a33279bdf194bd3ec4837f2737f63bcd0cb1c55a13429919" exitCode=0 Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.612826 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0b98e7b2-3068-4f59-b25f-4755bd752b49","Type":"ContainerDied","Data":"960d7fcb921daf08a33279bdf194bd3ec4837f2737f63bcd0cb1c55a13429919"} Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.616932 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gp6dc" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.617755 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gp6dc" event={"ID":"a24929f6-66ff-415c-b539-a9e6ac507f1d","Type":"ContainerDied","Data":"ffe481017dd6db707c56b4e05c0ebffff5f260e6c7ae470379fbfc924033c4ef"} Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.617894 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffe481017dd6db707c56b4e05c0ebffff5f260e6c7ae470379fbfc924033c4ef" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.619474 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e541-account-create-update-7769t" event={"ID":"84c1760e-09f1-4f63-ae74-8c8ba4af49be","Type":"ContainerDied","Data":"ae55c307c6d91af918bf42b4405a4c32a40efb4d0094a9218c94cd9d7139388c"} Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.619555 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae55c307c6d91af918bf42b4405a4c32a40efb4d0094a9218c94cd9d7139388c" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.619672 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e541-account-create-update-7769t" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.628569 4815 generic.go:334] "Generic (PLEG): container finished" podID="3ade0feb-5e78-4cf2-a82e-bad0e5468bc7" containerID="0fdd4260540bf1feb5f5070423d97484add03687a692d07a94d34439324625a7" exitCode=0 Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.628624 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zplf2" event={"ID":"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7","Type":"ContainerDied","Data":"0fdd4260540bf1feb5f5070423d97484add03687a692d07a94d34439324625a7"} Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.635256 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-29cc-account-create-update-rf4f5" event={"ID":"64ed35f4-2af0-4d3e-84a7-6160ac5b9185","Type":"ContainerDied","Data":"2c26ffd234dcc02874056d518d84acea0a897401521108b6c5ea888b6c1654b4"} Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.635296 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c26ffd234dcc02874056d518d84acea0a897401521108b6c5ea888b6c1654b4" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.635353 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-29cc-account-create-update-rf4f5" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.640566 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-4xctv" event={"ID":"50e8c23e-f13a-4fb9-b223-9330b3d1d36b","Type":"ContainerDied","Data":"018f5d5f03ccd378da703951e73bb703ae373c1b137d005da247565cbfbbee5b"} Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.640599 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="018f5d5f03ccd378da703951e73bb703ae373c1b137d005da247565cbfbbee5b" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.640684 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-4xctv" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.700995 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhlv7\" (UniqueName: \"kubernetes.io/projected/298a9862-55ab-4180-b10f-375379a96a2b-kube-api-access-mhlv7\") pod \"glance-df70-account-create-update-s4qnr\" (UID: \"298a9862-55ab-4180-b10f-375379a96a2b\") " pod="openstack/glance-df70-account-create-update-s4qnr" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.701055 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80276532-5b43-42b3-86db-de225388a3db-operator-scripts\") pod \"glance-db-create-st8v2\" (UID: \"80276532-5b43-42b3-86db-de225388a3db\") " pod="openstack/glance-db-create-st8v2" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.701178 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/298a9862-55ab-4180-b10f-375379a96a2b-operator-scripts\") pod \"glance-df70-account-create-update-s4qnr\" (UID: \"298a9862-55ab-4180-b10f-375379a96a2b\") " pod="openstack/glance-df70-account-create-update-s4qnr" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.701212 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfgzf\" (UniqueName: \"kubernetes.io/projected/80276532-5b43-42b3-86db-de225388a3db-kube-api-access-qfgzf\") pod \"glance-db-create-st8v2\" (UID: \"80276532-5b43-42b3-86db-de225388a3db\") " pod="openstack/glance-db-create-st8v2" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.702414 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80276532-5b43-42b3-86db-de225388a3db-operator-scripts\") pod \"glance-db-create-st8v2\" (UID: \"80276532-5b43-42b3-86db-de225388a3db\") " pod="openstack/glance-db-create-st8v2" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.716891 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfgzf\" (UniqueName: \"kubernetes.io/projected/80276532-5b43-42b3-86db-de225388a3db-kube-api-access-qfgzf\") pod \"glance-db-create-st8v2\" (UID: \"80276532-5b43-42b3-86db-de225388a3db\") " pod="openstack/glance-db-create-st8v2" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.805365 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhlv7\" (UniqueName: \"kubernetes.io/projected/298a9862-55ab-4180-b10f-375379a96a2b-kube-api-access-mhlv7\") pod \"glance-df70-account-create-update-s4qnr\" (UID: \"298a9862-55ab-4180-b10f-375379a96a2b\") " pod="openstack/glance-df70-account-create-update-s4qnr" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.805585 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/298a9862-55ab-4180-b10f-375379a96a2b-operator-scripts\") pod \"glance-df70-account-create-update-s4qnr\" (UID: \"298a9862-55ab-4180-b10f-375379a96a2b\") " pod="openstack/glance-df70-account-create-update-s4qnr" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.806482 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/298a9862-55ab-4180-b10f-375379a96a2b-operator-scripts\") pod \"glance-df70-account-create-update-s4qnr\" (UID: \"298a9862-55ab-4180-b10f-375379a96a2b\") " pod="openstack/glance-df70-account-create-update-s4qnr" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.808094 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-st8v2" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.845677 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhlv7\" (UniqueName: \"kubernetes.io/projected/298a9862-55ab-4180-b10f-375379a96a2b-kube-api-access-mhlv7\") pod \"glance-df70-account-create-update-s4qnr\" (UID: \"298a9862-55ab-4180-b10f-375379a96a2b\") " pod="openstack/glance-df70-account-create-update-s4qnr" Feb 25 13:36:55 crc kubenswrapper[4815]: I0225 13:36:55.905979 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-df70-account-create-update-s4qnr" Feb 25 13:36:56 crc kubenswrapper[4815]: I0225 13:36:56.260755 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 25 13:36:56 crc kubenswrapper[4815]: I0225 13:36:56.354312 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-st8v2"] Feb 25 13:36:56 crc kubenswrapper[4815]: I0225 13:36:56.521342 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-df70-account-create-update-s4qnr"] Feb 25 13:36:56 crc kubenswrapper[4815]: W0225 13:36:56.528176 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod298a9862_55ab_4180_b10f_375379a96a2b.slice/crio-312c41e947ae592fa12a03cbaddc220d9c398c73c9c79a0aff521fc8a218ab82 WatchSource:0}: Error finding container 312c41e947ae592fa12a03cbaddc220d9c398c73c9c79a0aff521fc8a218ab82: Status 404 returned error can't find the container with id 312c41e947ae592fa12a03cbaddc220d9c398c73c9c79a0aff521fc8a218ab82 Feb 25 13:36:56 crc kubenswrapper[4815]: I0225 13:36:56.648765 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-df70-account-create-update-s4qnr" event={"ID":"298a9862-55ab-4180-b10f-375379a96a2b","Type":"ContainerStarted","Data":"312c41e947ae592fa12a03cbaddc220d9c398c73c9c79a0aff521fc8a218ab82"} Feb 25 13:36:56 crc kubenswrapper[4815]: I0225 13:36:56.651604 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0b98e7b2-3068-4f59-b25f-4755bd752b49","Type":"ContainerStarted","Data":"9abd009533a8d747d7e9365f84325c58794cd02960b609440095ab81603d616c"} Feb 25 13:36:56 crc kubenswrapper[4815]: I0225 13:36:56.651828 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:36:56 crc kubenswrapper[4815]: I0225 13:36:56.653241 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-st8v2" event={"ID":"80276532-5b43-42b3-86db-de225388a3db","Type":"ContainerStarted","Data":"0795609e27a96f4062f243e6c5401f60a1d133919c6b727cfab7be2b687b257b"} Feb 25 13:36:56 crc kubenswrapper[4815]: I0225 13:36:56.653272 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-st8v2" event={"ID":"80276532-5b43-42b3-86db-de225388a3db","Type":"ContainerStarted","Data":"b3cd8c450a3ca83e06ad37f15e9da19323d8bdf02bbd52c2786665ecb8e87e6b"} Feb 25 13:36:56 crc kubenswrapper[4815]: I0225 13:36:56.686866 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.554285608 podStartE2EDuration="50.686845644s" podCreationTimestamp="2026-02-25 13:36:06 +0000 UTC" firstStartedPulling="2026-02-25 13:36:08.104429387 +0000 UTC m=+925.905527441" lastFinishedPulling="2026-02-25 13:36:22.236989413 +0000 UTC m=+940.038087477" observedRunningTime="2026-02-25 13:36:56.679752004 +0000 UTC m=+974.480850058" watchObservedRunningTime="2026-02-25 13:36:56.686845644 +0000 UTC m=+974.487943698" Feb 25 13:36:56 crc kubenswrapper[4815]: I0225 13:36:56.699126 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-st8v2" podStartSLOduration=1.699104072 podStartE2EDuration="1.699104072s" podCreationTimestamp="2026-02-25 13:36:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:36:56.696256344 +0000 UTC m=+974.497354398" watchObservedRunningTime="2026-02-25 13:36:56.699104072 +0000 UTC m=+974.500202126" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.066139 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.188681 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-rkpgf"] Feb 25 13:36:57 crc kubenswrapper[4815]: E0225 13:36:57.189341 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ade0feb-5e78-4cf2-a82e-bad0e5468bc7" containerName="swift-ring-rebalance" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.189365 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ade0feb-5e78-4cf2-a82e-bad0e5468bc7" containerName="swift-ring-rebalance" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.189590 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ade0feb-5e78-4cf2-a82e-bad0e5468bc7" containerName="swift-ring-rebalance" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.190148 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rkpgf" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.193130 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.203916 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-rkpgf"] Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.233114 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-etc-swift\") pod \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.233200 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-dispersionconf\") pod \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.233260 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-ring-data-devices\") pod \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.233319 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-swiftconf\") pod \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.233376 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-scripts\") pod \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.233398 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fvcr\" (UniqueName: \"kubernetes.io/projected/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-kube-api-access-9fvcr\") pod \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.233546 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-combined-ca-bundle\") pod \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\" (UID: \"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7\") " Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.234612 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3ade0feb-5e78-4cf2-a82e-bad0e5468bc7" (UID: "3ade0feb-5e78-4cf2-a82e-bad0e5468bc7"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.236720 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3ade0feb-5e78-4cf2-a82e-bad0e5468bc7" (UID: "3ade0feb-5e78-4cf2-a82e-bad0e5468bc7"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.261328 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-kube-api-access-9fvcr" (OuterVolumeSpecName: "kube-api-access-9fvcr") pod "3ade0feb-5e78-4cf2-a82e-bad0e5468bc7" (UID: "3ade0feb-5e78-4cf2-a82e-bad0e5468bc7"). InnerVolumeSpecName "kube-api-access-9fvcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.264347 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3ade0feb-5e78-4cf2-a82e-bad0e5468bc7" (UID: "3ade0feb-5e78-4cf2-a82e-bad0e5468bc7"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.273671 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3ade0feb-5e78-4cf2-a82e-bad0e5468bc7" (UID: "3ade0feb-5e78-4cf2-a82e-bad0e5468bc7"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.275366 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ade0feb-5e78-4cf2-a82e-bad0e5468bc7" (UID: "3ade0feb-5e78-4cf2-a82e-bad0e5468bc7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.291674 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-scripts" (OuterVolumeSpecName: "scripts") pod "3ade0feb-5e78-4cf2-a82e-bad0e5468bc7" (UID: "3ade0feb-5e78-4cf2-a82e-bad0e5468bc7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.335488 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4bd962e-a6c4-4d4d-8b69-f4fff5af482c-operator-scripts\") pod \"root-account-create-update-rkpgf\" (UID: \"c4bd962e-a6c4-4d4d-8b69-f4fff5af482c\") " pod="openstack/root-account-create-update-rkpgf" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.335718 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfpss\" (UniqueName: \"kubernetes.io/projected/c4bd962e-a6c4-4d4d-8b69-f4fff5af482c-kube-api-access-xfpss\") pod \"root-account-create-update-rkpgf\" (UID: \"c4bd962e-a6c4-4d4d-8b69-f4fff5af482c\") " pod="openstack/root-account-create-update-rkpgf" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.335773 4815 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.335786 4815 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.335795 4815 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.335807 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.335820 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fvcr\" (UniqueName: \"kubernetes.io/projected/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-kube-api-access-9fvcr\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.335835 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.335847 4815 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ade0feb-5e78-4cf2-a82e-bad0e5468bc7-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.437318 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfpss\" (UniqueName: \"kubernetes.io/projected/c4bd962e-a6c4-4d4d-8b69-f4fff5af482c-kube-api-access-xfpss\") pod \"root-account-create-update-rkpgf\" (UID: \"c4bd962e-a6c4-4d4d-8b69-f4fff5af482c\") " pod="openstack/root-account-create-update-rkpgf" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.437385 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4bd962e-a6c4-4d4d-8b69-f4fff5af482c-operator-scripts\") pod \"root-account-create-update-rkpgf\" (UID: \"c4bd962e-a6c4-4d4d-8b69-f4fff5af482c\") " pod="openstack/root-account-create-update-rkpgf" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.438079 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4bd962e-a6c4-4d4d-8b69-f4fff5af482c-operator-scripts\") pod \"root-account-create-update-rkpgf\" (UID: \"c4bd962e-a6c4-4d4d-8b69-f4fff5af482c\") " pod="openstack/root-account-create-update-rkpgf" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.452127 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfpss\" (UniqueName: \"kubernetes.io/projected/c4bd962e-a6c4-4d4d-8b69-f4fff5af482c-kube-api-access-xfpss\") pod \"root-account-create-update-rkpgf\" (UID: \"c4bd962e-a6c4-4d4d-8b69-f4fff5af482c\") " pod="openstack/root-account-create-update-rkpgf" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.510034 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rkpgf" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.663616 4815 generic.go:334] "Generic (PLEG): container finished" podID="298a9862-55ab-4180-b10f-375379a96a2b" containerID="96f99d91e60be382a1ff85ca50faf71803c3409dd984d5cb89b984c01d94cf00" exitCode=0 Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.663704 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-df70-account-create-update-s4qnr" event={"ID":"298a9862-55ab-4180-b10f-375379a96a2b","Type":"ContainerDied","Data":"96f99d91e60be382a1ff85ca50faf71803c3409dd984d5cb89b984c01d94cf00"} Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.666791 4815 generic.go:334] "Generic (PLEG): container finished" podID="80276532-5b43-42b3-86db-de225388a3db" containerID="0795609e27a96f4062f243e6c5401f60a1d133919c6b727cfab7be2b687b257b" exitCode=0 Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.666865 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-st8v2" event={"ID":"80276532-5b43-42b3-86db-de225388a3db","Type":"ContainerDied","Data":"0795609e27a96f4062f243e6c5401f60a1d133919c6b727cfab7be2b687b257b"} Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.677328 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zplf2" Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.679677 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zplf2" event={"ID":"3ade0feb-5e78-4cf2-a82e-bad0e5468bc7","Type":"ContainerDied","Data":"860d930d6766f6f79181fd8d9396f3b1385e634c4154223530dfcf1d9a41cfb7"} Feb 25 13:36:57 crc kubenswrapper[4815]: I0225 13:36:57.679714 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="860d930d6766f6f79181fd8d9396f3b1385e634c4154223530dfcf1d9a41cfb7" Feb 25 13:36:58 crc kubenswrapper[4815]: I0225 13:36:58.001744 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-rkpgf"] Feb 25 13:36:58 crc kubenswrapper[4815]: W0225 13:36:58.005477 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4bd962e_a6c4_4d4d_8b69_f4fff5af482c.slice/crio-000c1c4886b30a416f5527d1e4e20ff409a8e5bc2294c9e754374673bab695de WatchSource:0}: Error finding container 000c1c4886b30a416f5527d1e4e20ff409a8e5bc2294c9e754374673bab695de: Status 404 returned error can't find the container with id 000c1c4886b30a416f5527d1e4e20ff409a8e5bc2294c9e754374673bab695de Feb 25 13:36:58 crc kubenswrapper[4815]: I0225 13:36:58.685522 4815 generic.go:334] "Generic (PLEG): container finished" podID="c4bd962e-a6c4-4d4d-8b69-f4fff5af482c" containerID="9ce8c60b92b710cea3afd50b2e1693cb413c6423d29d6fa7e857c9247b4c3947" exitCode=0 Feb 25 13:36:58 crc kubenswrapper[4815]: I0225 13:36:58.685623 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rkpgf" event={"ID":"c4bd962e-a6c4-4d4d-8b69-f4fff5af482c","Type":"ContainerDied","Data":"9ce8c60b92b710cea3afd50b2e1693cb413c6423d29d6fa7e857c9247b4c3947"} Feb 25 13:36:58 crc kubenswrapper[4815]: I0225 13:36:58.685897 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rkpgf" event={"ID":"c4bd962e-a6c4-4d4d-8b69-f4fff5af482c","Type":"ContainerStarted","Data":"000c1c4886b30a416f5527d1e4e20ff409a8e5bc2294c9e754374673bab695de"} Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.196419 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-st8v2" Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.200523 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-df70-account-create-update-s4qnr" Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.377157 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhlv7\" (UniqueName: \"kubernetes.io/projected/298a9862-55ab-4180-b10f-375379a96a2b-kube-api-access-mhlv7\") pod \"298a9862-55ab-4180-b10f-375379a96a2b\" (UID: \"298a9862-55ab-4180-b10f-375379a96a2b\") " Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.377207 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80276532-5b43-42b3-86db-de225388a3db-operator-scripts\") pod \"80276532-5b43-42b3-86db-de225388a3db\" (UID: \"80276532-5b43-42b3-86db-de225388a3db\") " Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.377285 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/298a9862-55ab-4180-b10f-375379a96a2b-operator-scripts\") pod \"298a9862-55ab-4180-b10f-375379a96a2b\" (UID: \"298a9862-55ab-4180-b10f-375379a96a2b\") " Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.377300 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfgzf\" (UniqueName: \"kubernetes.io/projected/80276532-5b43-42b3-86db-de225388a3db-kube-api-access-qfgzf\") pod \"80276532-5b43-42b3-86db-de225388a3db\" (UID: \"80276532-5b43-42b3-86db-de225388a3db\") " Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.378900 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80276532-5b43-42b3-86db-de225388a3db-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "80276532-5b43-42b3-86db-de225388a3db" (UID: "80276532-5b43-42b3-86db-de225388a3db"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.378910 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/298a9862-55ab-4180-b10f-375379a96a2b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "298a9862-55ab-4180-b10f-375379a96a2b" (UID: "298a9862-55ab-4180-b10f-375379a96a2b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.387730 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80276532-5b43-42b3-86db-de225388a3db-kube-api-access-qfgzf" (OuterVolumeSpecName: "kube-api-access-qfgzf") pod "80276532-5b43-42b3-86db-de225388a3db" (UID: "80276532-5b43-42b3-86db-de225388a3db"). InnerVolumeSpecName "kube-api-access-qfgzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.398732 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/298a9862-55ab-4180-b10f-375379a96a2b-kube-api-access-mhlv7" (OuterVolumeSpecName: "kube-api-access-mhlv7") pod "298a9862-55ab-4180-b10f-375379a96a2b" (UID: "298a9862-55ab-4180-b10f-375379a96a2b"). InnerVolumeSpecName "kube-api-access-mhlv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.479131 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhlv7\" (UniqueName: \"kubernetes.io/projected/298a9862-55ab-4180-b10f-375379a96a2b-kube-api-access-mhlv7\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.479167 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80276532-5b43-42b3-86db-de225388a3db-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.479180 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/298a9862-55ab-4180-b10f-375379a96a2b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.479191 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfgzf\" (UniqueName: \"kubernetes.io/projected/80276532-5b43-42b3-86db-de225388a3db-kube-api-access-qfgzf\") on node \"crc\" DevicePath \"\"" Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.695877 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-df70-account-create-update-s4qnr" event={"ID":"298a9862-55ab-4180-b10f-375379a96a2b","Type":"ContainerDied","Data":"312c41e947ae592fa12a03cbaddc220d9c398c73c9c79a0aff521fc8a218ab82"} Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.695926 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="312c41e947ae592fa12a03cbaddc220d9c398c73c9c79a0aff521fc8a218ab82" Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.695890 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-df70-account-create-update-s4qnr" Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.697618 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-st8v2" event={"ID":"80276532-5b43-42b3-86db-de225388a3db","Type":"ContainerDied","Data":"b3cd8c450a3ca83e06ad37f15e9da19323d8bdf02bbd52c2786665ecb8e87e6b"} Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.697651 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3cd8c450a3ca83e06ad37f15e9da19323d8bdf02bbd52c2786665ecb8e87e6b" Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.697662 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-st8v2" Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.987372 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:36:59 crc kubenswrapper[4815]: I0225 13:36:59.992412 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3941650f-c99b-4043-9137-e4c43361f93f-etc-swift\") pod \"swift-storage-0\" (UID: \"3941650f-c99b-4043-9137-e4c43361f93f\") " pod="openstack/swift-storage-0" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.019943 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.121085 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rkpgf" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.190115 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfpss\" (UniqueName: \"kubernetes.io/projected/c4bd962e-a6c4-4d4d-8b69-f4fff5af482c-kube-api-access-xfpss\") pod \"c4bd962e-a6c4-4d4d-8b69-f4fff5af482c\" (UID: \"c4bd962e-a6c4-4d4d-8b69-f4fff5af482c\") " Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.190212 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4bd962e-a6c4-4d4d-8b69-f4fff5af482c-operator-scripts\") pod \"c4bd962e-a6c4-4d4d-8b69-f4fff5af482c\" (UID: \"c4bd962e-a6c4-4d4d-8b69-f4fff5af482c\") " Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.190640 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4bd962e-a6c4-4d4d-8b69-f4fff5af482c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c4bd962e-a6c4-4d4d-8b69-f4fff5af482c" (UID: "c4bd962e-a6c4-4d4d-8b69-f4fff5af482c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.200755 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4bd962e-a6c4-4d4d-8b69-f4fff5af482c-kube-api-access-xfpss" (OuterVolumeSpecName: "kube-api-access-xfpss") pod "c4bd962e-a6c4-4d4d-8b69-f4fff5af482c" (UID: "c4bd962e-a6c4-4d4d-8b69-f4fff5af482c"). InnerVolumeSpecName "kube-api-access-xfpss". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.291597 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4bd962e-a6c4-4d4d-8b69-f4fff5af482c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.291638 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfpss\" (UniqueName: \"kubernetes.io/projected/c4bd962e-a6c4-4d4d-8b69-f4fff5af482c-kube-api-access-xfpss\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.607126 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 25 13:37:00 crc kubenswrapper[4815]: W0225 13:37:00.609751 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3941650f_c99b_4043_9137_e4c43361f93f.slice/crio-85724cfef5219ca5a7d01d053aef5c0ff4aeca8634d4c3f8a20464b21e8b8f32 WatchSource:0}: Error finding container 85724cfef5219ca5a7d01d053aef5c0ff4aeca8634d4c3f8a20464b21e8b8f32: Status 404 returned error can't find the container with id 85724cfef5219ca5a7d01d053aef5c0ff4aeca8634d4c3f8a20464b21e8b8f32 Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.704579 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3941650f-c99b-4043-9137-e4c43361f93f","Type":"ContainerStarted","Data":"85724cfef5219ca5a7d01d053aef5c0ff4aeca8634d4c3f8a20464b21e8b8f32"} Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.705888 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rkpgf" event={"ID":"c4bd962e-a6c4-4d4d-8b69-f4fff5af482c","Type":"ContainerDied","Data":"000c1c4886b30a416f5527d1e4e20ff409a8e5bc2294c9e754374673bab695de"} Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.705918 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="000c1c4886b30a416f5527d1e4e20ff409a8e5bc2294c9e754374673bab695de" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.705962 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rkpgf" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.826637 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-b9qrv"] Feb 25 13:37:00 crc kubenswrapper[4815]: E0225 13:37:00.827191 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80276532-5b43-42b3-86db-de225388a3db" containerName="mariadb-database-create" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.827207 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="80276532-5b43-42b3-86db-de225388a3db" containerName="mariadb-database-create" Feb 25 13:37:00 crc kubenswrapper[4815]: E0225 13:37:00.827227 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="298a9862-55ab-4180-b10f-375379a96a2b" containerName="mariadb-account-create-update" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.827234 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="298a9862-55ab-4180-b10f-375379a96a2b" containerName="mariadb-account-create-update" Feb 25 13:37:00 crc kubenswrapper[4815]: E0225 13:37:00.827247 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4bd962e-a6c4-4d4d-8b69-f4fff5af482c" containerName="mariadb-account-create-update" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.827254 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4bd962e-a6c4-4d4d-8b69-f4fff5af482c" containerName="mariadb-account-create-update" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.827411 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4bd962e-a6c4-4d4d-8b69-f4fff5af482c" containerName="mariadb-account-create-update" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.827423 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="298a9862-55ab-4180-b10f-375379a96a2b" containerName="mariadb-account-create-update" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.827435 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="80276532-5b43-42b3-86db-de225388a3db" containerName="mariadb-database-create" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.827919 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-b9qrv" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.829905 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.834028 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-dqnk7" Feb 25 13:37:00 crc kubenswrapper[4815]: I0225 13:37:00.836105 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-b9qrv"] Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.002126 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxvhb\" (UniqueName: \"kubernetes.io/projected/7e9844cb-1b51-4e42-8fee-5ed7059776c1-kube-api-access-cxvhb\") pod \"glance-db-sync-b9qrv\" (UID: \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\") " pod="openstack/glance-db-sync-b9qrv" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.002332 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-config-data\") pod \"glance-db-sync-b9qrv\" (UID: \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\") " pod="openstack/glance-db-sync-b9qrv" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.002435 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-combined-ca-bundle\") pod \"glance-db-sync-b9qrv\" (UID: \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\") " pod="openstack/glance-db-sync-b9qrv" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.002540 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-db-sync-config-data\") pod \"glance-db-sync-b9qrv\" (UID: \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\") " pod="openstack/glance-db-sync-b9qrv" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.104071 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxvhb\" (UniqueName: \"kubernetes.io/projected/7e9844cb-1b51-4e42-8fee-5ed7059776c1-kube-api-access-cxvhb\") pod \"glance-db-sync-b9qrv\" (UID: \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\") " pod="openstack/glance-db-sync-b9qrv" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.104126 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-config-data\") pod \"glance-db-sync-b9qrv\" (UID: \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\") " pod="openstack/glance-db-sync-b9qrv" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.104172 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-combined-ca-bundle\") pod \"glance-db-sync-b9qrv\" (UID: \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\") " pod="openstack/glance-db-sync-b9qrv" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.104222 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-db-sync-config-data\") pod \"glance-db-sync-b9qrv\" (UID: \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\") " pod="openstack/glance-db-sync-b9qrv" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.109080 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-combined-ca-bundle\") pod \"glance-db-sync-b9qrv\" (UID: \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\") " pod="openstack/glance-db-sync-b9qrv" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.109229 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-config-data\") pod \"glance-db-sync-b9qrv\" (UID: \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\") " pod="openstack/glance-db-sync-b9qrv" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.109487 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-db-sync-config-data\") pod \"glance-db-sync-b9qrv\" (UID: \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\") " pod="openstack/glance-db-sync-b9qrv" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.124477 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxvhb\" (UniqueName: \"kubernetes.io/projected/7e9844cb-1b51-4e42-8fee-5ed7059776c1-kube-api-access-cxvhb\") pod \"glance-db-sync-b9qrv\" (UID: \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\") " pod="openstack/glance-db-sync-b9qrv" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.148157 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-b9qrv" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.316819 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-tjrr6" podUID="d9d08116-2a87-4ae4-8485-dab2bb05de4c" containerName="ovn-controller" probeResult="failure" output=< Feb 25 13:37:01 crc kubenswrapper[4815]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 25 13:37:01 crc kubenswrapper[4815]: > Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.328762 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.331231 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-z74w4" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.543665 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tjrr6-config-ptp2v"] Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.545798 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.548161 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.561743 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tjrr6-config-ptp2v"] Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.663389 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-b9qrv"] Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.712536 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-scripts\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.712646 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-run-ovn\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.712695 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pdfx\" (UniqueName: \"kubernetes.io/projected/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-kube-api-access-4pdfx\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.712724 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-additional-scripts\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.712758 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-log-ovn\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.713430 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-run\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.814846 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-run-ovn\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.814916 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pdfx\" (UniqueName: \"kubernetes.io/projected/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-kube-api-access-4pdfx\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.814993 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-additional-scripts\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.815043 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-log-ovn\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.815126 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-run\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.815227 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-scripts\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.815672 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-run-ovn\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.815702 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-log-ovn\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.815765 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-run\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.816329 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-additional-scripts\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.818460 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-scripts\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.838728 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pdfx\" (UniqueName: \"kubernetes.io/projected/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-kube-api-access-4pdfx\") pod \"ovn-controller-tjrr6-config-ptp2v\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:01 crc kubenswrapper[4815]: I0225 13:37:01.863522 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:02 crc kubenswrapper[4815]: I0225 13:37:02.385649 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tjrr6-config-ptp2v"] Feb 25 13:37:02 crc kubenswrapper[4815]: W0225 13:37:02.408329 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78ad93de_f5cd_4fe2_8bbf_a75f36aec953.slice/crio-15355479318deea27190d6eb078d38c183df9e90da329261a16637249d7fe53d WatchSource:0}: Error finding container 15355479318deea27190d6eb078d38c183df9e90da329261a16637249d7fe53d: Status 404 returned error can't find the container with id 15355479318deea27190d6eb078d38c183df9e90da329261a16637249d7fe53d Feb 25 13:37:02 crc kubenswrapper[4815]: I0225 13:37:02.723222 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3941650f-c99b-4043-9137-e4c43361f93f","Type":"ContainerStarted","Data":"48e9be6d887eb802e2a499d5df6102c34e6462d0caf04b1990eb8649a3d1d76c"} Feb 25 13:37:02 crc kubenswrapper[4815]: I0225 13:37:02.723545 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3941650f-c99b-4043-9137-e4c43361f93f","Type":"ContainerStarted","Data":"7a06dafe1e920e36b21256f926b8a9b2782e15b095823375ee9386c44da4a257"} Feb 25 13:37:02 crc kubenswrapper[4815]: I0225 13:37:02.723557 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3941650f-c99b-4043-9137-e4c43361f93f","Type":"ContainerStarted","Data":"686d81e79399a6db905a50c3b4d4c2221f594e1f10b9f10e8f14b99254838b18"} Feb 25 13:37:02 crc kubenswrapper[4815]: I0225 13:37:02.723565 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3941650f-c99b-4043-9137-e4c43361f93f","Type":"ContainerStarted","Data":"d41960d32e6e378d2dadce315d4c9538bd71e1d4bdaa70d10285bf9b7671c256"} Feb 25 13:37:02 crc kubenswrapper[4815]: I0225 13:37:02.724707 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-b9qrv" event={"ID":"7e9844cb-1b51-4e42-8fee-5ed7059776c1","Type":"ContainerStarted","Data":"de6a14523ff7330a0720239d4ab5eb7b7d52576f29f265da5a7d99b2090f91c3"} Feb 25 13:37:02 crc kubenswrapper[4815]: I0225 13:37:02.726721 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tjrr6-config-ptp2v" event={"ID":"78ad93de-f5cd-4fe2-8bbf-a75f36aec953","Type":"ContainerStarted","Data":"15355479318deea27190d6eb078d38c183df9e90da329261a16637249d7fe53d"} Feb 25 13:37:03 crc kubenswrapper[4815]: I0225 13:37:03.743562 4815 generic.go:334] "Generic (PLEG): container finished" podID="78ad93de-f5cd-4fe2-8bbf-a75f36aec953" containerID="ebe6f94cd2653d2dbce26684653779141960ca6b8db847bde41f11f2653d6590" exitCode=0 Feb 25 13:37:03 crc kubenswrapper[4815]: I0225 13:37:03.743904 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tjrr6-config-ptp2v" event={"ID":"78ad93de-f5cd-4fe2-8bbf-a75f36aec953","Type":"ContainerDied","Data":"ebe6f94cd2653d2dbce26684653779141960ca6b8db847bde41f11f2653d6590"} Feb 25 13:37:03 crc kubenswrapper[4815]: I0225 13:37:03.775408 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-rkpgf"] Feb 25 13:37:03 crc kubenswrapper[4815]: I0225 13:37:03.793738 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-rkpgf"] Feb 25 13:37:04 crc kubenswrapper[4815]: I0225 13:37:04.759678 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3941650f-c99b-4043-9137-e4c43361f93f","Type":"ContainerStarted","Data":"aac62bb59bc2f4c14bde88f94cffb74ca33e251839b338c8e82823148173e5f9"} Feb 25 13:37:04 crc kubenswrapper[4815]: I0225 13:37:04.760000 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3941650f-c99b-4043-9137-e4c43361f93f","Type":"ContainerStarted","Data":"273cd8c70a542a8fc004d5e9ce03ae3133cfc593aa7469fff84e4bc421ee312f"} Feb 25 13:37:04 crc kubenswrapper[4815]: I0225 13:37:04.948633 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4bd962e-a6c4-4d4d-8b69-f4fff5af482c" path="/var/lib/kubelet/pods/c4bd962e-a6c4-4d4d-8b69-f4fff5af482c/volumes" Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.072834 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.186259 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-run\") pod \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.186375 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-run" (OuterVolumeSpecName: "var-run") pod "78ad93de-f5cd-4fe2-8bbf-a75f36aec953" (UID: "78ad93de-f5cd-4fe2-8bbf-a75f36aec953"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.186443 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-additional-scripts\") pod \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.186538 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-run-ovn\") pod \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.186600 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pdfx\" (UniqueName: \"kubernetes.io/projected/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-kube-api-access-4pdfx\") pod \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.186648 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-log-ovn\") pod \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.186662 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "78ad93de-f5cd-4fe2-8bbf-a75f36aec953" (UID: "78ad93de-f5cd-4fe2-8bbf-a75f36aec953"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.187888 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "78ad93de-f5cd-4fe2-8bbf-a75f36aec953" (UID: "78ad93de-f5cd-4fe2-8bbf-a75f36aec953"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.189630 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "78ad93de-f5cd-4fe2-8bbf-a75f36aec953" (UID: "78ad93de-f5cd-4fe2-8bbf-a75f36aec953"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.190660 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-scripts" (OuterVolumeSpecName: "scripts") pod "78ad93de-f5cd-4fe2-8bbf-a75f36aec953" (UID: "78ad93de-f5cd-4fe2-8bbf-a75f36aec953"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.192158 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-kube-api-access-4pdfx" (OuterVolumeSpecName: "kube-api-access-4pdfx") pod "78ad93de-f5cd-4fe2-8bbf-a75f36aec953" (UID: "78ad93de-f5cd-4fe2-8bbf-a75f36aec953"). InnerVolumeSpecName "kube-api-access-4pdfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.186701 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-scripts\") pod \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\" (UID: \"78ad93de-f5cd-4fe2-8bbf-a75f36aec953\") " Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.193745 4815 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-run\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.193762 4815 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.193781 4815 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.193795 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pdfx\" (UniqueName: \"kubernetes.io/projected/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-kube-api-access-4pdfx\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.193807 4815 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.193820 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78ad93de-f5cd-4fe2-8bbf-a75f36aec953-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.777619 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3941650f-c99b-4043-9137-e4c43361f93f","Type":"ContainerStarted","Data":"03005a0680331f8722f584e831faee51a882ee595ab6a28bdef734edf67b0e7a"} Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.778495 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3941650f-c99b-4043-9137-e4c43361f93f","Type":"ContainerStarted","Data":"7c03d574143b3e31099f1338a0a5e1be39b63c4203a065a46f24225198badbdd"} Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.780163 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tjrr6-config-ptp2v" event={"ID":"78ad93de-f5cd-4fe2-8bbf-a75f36aec953","Type":"ContainerDied","Data":"15355479318deea27190d6eb078d38c183df9e90da329261a16637249d7fe53d"} Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.780263 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15355479318deea27190d6eb078d38c183df9e90da329261a16637249d7fe53d" Feb 25 13:37:05 crc kubenswrapper[4815]: I0225 13:37:05.780397 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tjrr6-config-ptp2v" Feb 25 13:37:06 crc kubenswrapper[4815]: I0225 13:37:06.187707 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-tjrr6-config-ptp2v"] Feb 25 13:37:06 crc kubenswrapper[4815]: I0225 13:37:06.201566 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-tjrr6-config-ptp2v"] Feb 25 13:37:06 crc kubenswrapper[4815]: I0225 13:37:06.302875 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-tjrr6" Feb 25 13:37:06 crc kubenswrapper[4815]: I0225 13:37:06.803290 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3941650f-c99b-4043-9137-e4c43361f93f","Type":"ContainerStarted","Data":"92d74f3cadb3472cbbc44cb34e4fc502137853b6d646d3a7f1b45a6b9625ce14"} Feb 25 13:37:06 crc kubenswrapper[4815]: I0225 13:37:06.803331 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3941650f-c99b-4043-9137-e4c43361f93f","Type":"ContainerStarted","Data":"f4480d7638e16a245f9c04100c0c3719e601965f1b89639ca9272a3dd69c7ee8"} Feb 25 13:37:06 crc kubenswrapper[4815]: I0225 13:37:06.803339 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3941650f-c99b-4043-9137-e4c43361f93f","Type":"ContainerStarted","Data":"b00ddca497e6fca4fdd6c7e1130d77769ffc6c3080437a01c21e29e2f81b16d1"} Feb 25 13:37:06 crc kubenswrapper[4815]: I0225 13:37:06.946361 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78ad93de-f5cd-4fe2-8bbf-a75f36aec953" path="/var/lib/kubelet/pods/78ad93de-f5cd-4fe2-8bbf-a75f36aec953/volumes" Feb 25 13:37:07 crc kubenswrapper[4815]: I0225 13:37:07.611760 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:37:07 crc kubenswrapper[4815]: I0225 13:37:07.816158 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3941650f-c99b-4043-9137-e4c43361f93f","Type":"ContainerStarted","Data":"2825b2839f877ca7f4cf9e368c4e4e7e1f9f4db6236bb6f4db12a5431561f28e"} Feb 25 13:37:07 crc kubenswrapper[4815]: I0225 13:37:07.816204 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3941650f-c99b-4043-9137-e4c43361f93f","Type":"ContainerStarted","Data":"44e4b3f6288c06c86a04a4848ee5c4eb942f5cc122fb8b866b00f9f8a35102a8"} Feb 25 13:37:07 crc kubenswrapper[4815]: I0225 13:37:07.816216 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3941650f-c99b-4043-9137-e4c43361f93f","Type":"ContainerStarted","Data":"5d380fe4d042bedf6a70f1a5df7bfe659b230515d1b318f321be6b79235ace78"} Feb 25 13:37:07 crc kubenswrapper[4815]: I0225 13:37:07.816224 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3941650f-c99b-4043-9137-e4c43361f93f","Type":"ContainerStarted","Data":"3c5f95a2abffa627191025abbcbaa40a5230acb37941e4f5399f5f118c561404"} Feb 25 13:37:07 crc kubenswrapper[4815]: I0225 13:37:07.872155 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.506699707 podStartE2EDuration="25.872131756s" podCreationTimestamp="2026-02-25 13:36:42 +0000 UTC" firstStartedPulling="2026-02-25 13:37:00.611778941 +0000 UTC m=+978.412876995" lastFinishedPulling="2026-02-25 13:37:05.97721099 +0000 UTC m=+983.778309044" observedRunningTime="2026-02-25 13:37:07.862945742 +0000 UTC m=+985.664043816" watchObservedRunningTime="2026-02-25 13:37:07.872131756 +0000 UTC m=+985.673229810" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.151115 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-swzkm"] Feb 25 13:37:08 crc kubenswrapper[4815]: E0225 13:37:08.152527 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78ad93de-f5cd-4fe2-8bbf-a75f36aec953" containerName="ovn-config" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.152552 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="78ad93de-f5cd-4fe2-8bbf-a75f36aec953" containerName="ovn-config" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.152745 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="78ad93de-f5cd-4fe2-8bbf-a75f36aec953" containerName="ovn-config" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.153589 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.155393 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.168275 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-swzkm"] Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.253405 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.253473 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.253680 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.253759 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-config\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.253836 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x5kh\" (UniqueName: \"kubernetes.io/projected/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-kube-api-access-5x5kh\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.253904 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.355102 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-config\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.355595 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x5kh\" (UniqueName: \"kubernetes.io/projected/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-kube-api-access-5x5kh\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.355638 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.355712 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.355745 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.355777 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.355974 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-config\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.356497 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.356588 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.357107 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.357450 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.383687 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x5kh\" (UniqueName: \"kubernetes.io/projected/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-kube-api-access-5x5kh\") pod \"dnsmasq-dns-6d5b6d6b67-swzkm\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.512182 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.766381 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-49c8k"] Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.767596 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-49c8k" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.769529 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.776575 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-49c8k"] Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.826327 4815 generic.go:334] "Generic (PLEG): container finished" podID="4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" containerID="e633733b61b772b04a5fcbcaa5076b5a7353c6b7fd094307d1bc64bbb9741d6c" exitCode=0 Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.826422 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d","Type":"ContainerDied","Data":"e633733b61b772b04a5fcbcaa5076b5a7353c6b7fd094307d1bc64bbb9741d6c"} Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.864360 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wnld\" (UniqueName: \"kubernetes.io/projected/6dbd3d76-08d8-4798-91ab-2dc5afe1993f-kube-api-access-7wnld\") pod \"root-account-create-update-49c8k\" (UID: \"6dbd3d76-08d8-4798-91ab-2dc5afe1993f\") " pod="openstack/root-account-create-update-49c8k" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.864433 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dbd3d76-08d8-4798-91ab-2dc5afe1993f-operator-scripts\") pod \"root-account-create-update-49c8k\" (UID: \"6dbd3d76-08d8-4798-91ab-2dc5afe1993f\") " pod="openstack/root-account-create-update-49c8k" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.966637 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wnld\" (UniqueName: \"kubernetes.io/projected/6dbd3d76-08d8-4798-91ab-2dc5afe1993f-kube-api-access-7wnld\") pod \"root-account-create-update-49c8k\" (UID: \"6dbd3d76-08d8-4798-91ab-2dc5afe1993f\") " pod="openstack/root-account-create-update-49c8k" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.969538 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dbd3d76-08d8-4798-91ab-2dc5afe1993f-operator-scripts\") pod \"root-account-create-update-49c8k\" (UID: \"6dbd3d76-08d8-4798-91ab-2dc5afe1993f\") " pod="openstack/root-account-create-update-49c8k" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.970692 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dbd3d76-08d8-4798-91ab-2dc5afe1993f-operator-scripts\") pod \"root-account-create-update-49c8k\" (UID: \"6dbd3d76-08d8-4798-91ab-2dc5afe1993f\") " pod="openstack/root-account-create-update-49c8k" Feb 25 13:37:08 crc kubenswrapper[4815]: I0225 13:37:08.989738 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wnld\" (UniqueName: \"kubernetes.io/projected/6dbd3d76-08d8-4798-91ab-2dc5afe1993f-kube-api-access-7wnld\") pod \"root-account-create-update-49c8k\" (UID: \"6dbd3d76-08d8-4798-91ab-2dc5afe1993f\") " pod="openstack/root-account-create-update-49c8k" Feb 25 13:37:09 crc kubenswrapper[4815]: I0225 13:37:09.082106 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-49c8k" Feb 25 13:37:16 crc kubenswrapper[4815]: I0225 13:37:16.218799 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-49c8k"] Feb 25 13:37:16 crc kubenswrapper[4815]: W0225 13:37:16.242949 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dbd3d76_08d8_4798_91ab_2dc5afe1993f.slice/crio-9a3e4c6876284fd8aa7436efcc41991076c3d959e60643e420cfbe42d1a44ad5 WatchSource:0}: Error finding container 9a3e4c6876284fd8aa7436efcc41991076c3d959e60643e420cfbe42d1a44ad5: Status 404 returned error can't find the container with id 9a3e4c6876284fd8aa7436efcc41991076c3d959e60643e420cfbe42d1a44ad5 Feb 25 13:37:16 crc kubenswrapper[4815]: I0225 13:37:16.468036 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-swzkm"] Feb 25 13:37:16 crc kubenswrapper[4815]: W0225 13:37:16.477084 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05c3c613_64ee_4c4d_a7b0_16265fc7ba09.slice/crio-ce044b46fef43b20c233823110d650e5a6b5c48e7207609ed2dc874167a3735a WatchSource:0}: Error finding container ce044b46fef43b20c233823110d650e5a6b5c48e7207609ed2dc874167a3735a: Status 404 returned error can't find the container with id ce044b46fef43b20c233823110d650e5a6b5c48e7207609ed2dc874167a3735a Feb 25 13:37:16 crc kubenswrapper[4815]: I0225 13:37:16.908391 4815 generic.go:334] "Generic (PLEG): container finished" podID="6dbd3d76-08d8-4798-91ab-2dc5afe1993f" containerID="c8098a5c9f4fded8b1aa8dc1f4360ba63f0c8505eb27faffe83a00e7c0741271" exitCode=0 Feb 25 13:37:16 crc kubenswrapper[4815]: I0225 13:37:16.908470 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-49c8k" event={"ID":"6dbd3d76-08d8-4798-91ab-2dc5afe1993f","Type":"ContainerDied","Data":"c8098a5c9f4fded8b1aa8dc1f4360ba63f0c8505eb27faffe83a00e7c0741271"} Feb 25 13:37:16 crc kubenswrapper[4815]: I0225 13:37:16.908517 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-49c8k" event={"ID":"6dbd3d76-08d8-4798-91ab-2dc5afe1993f","Type":"ContainerStarted","Data":"9a3e4c6876284fd8aa7436efcc41991076c3d959e60643e420cfbe42d1a44ad5"} Feb 25 13:37:16 crc kubenswrapper[4815]: I0225 13:37:16.911805 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d","Type":"ContainerStarted","Data":"7e7cb8ec442c6d0abc60fe0d32bc0d3e4bb7cbc24ea616784c6b8eba67f393c7"} Feb 25 13:37:16 crc kubenswrapper[4815]: I0225 13:37:16.912118 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 25 13:37:16 crc kubenswrapper[4815]: I0225 13:37:16.914668 4815 generic.go:334] "Generic (PLEG): container finished" podID="05c3c613-64ee-4c4d-a7b0-16265fc7ba09" containerID="4ba253a99bac5cd3e4f35dc846673196ba5ef58e76e79061dabd4217083cee4d" exitCode=0 Feb 25 13:37:16 crc kubenswrapper[4815]: I0225 13:37:16.914744 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" event={"ID":"05c3c613-64ee-4c4d-a7b0-16265fc7ba09","Type":"ContainerDied","Data":"4ba253a99bac5cd3e4f35dc846673196ba5ef58e76e79061dabd4217083cee4d"} Feb 25 13:37:16 crc kubenswrapper[4815]: I0225 13:37:16.914775 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" event={"ID":"05c3c613-64ee-4c4d-a7b0-16265fc7ba09","Type":"ContainerStarted","Data":"ce044b46fef43b20c233823110d650e5a6b5c48e7207609ed2dc874167a3735a"} Feb 25 13:37:16 crc kubenswrapper[4815]: I0225 13:37:16.917213 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-b9qrv" event={"ID":"7e9844cb-1b51-4e42-8fee-5ed7059776c1","Type":"ContainerStarted","Data":"f2018e423eff7617abcb82018b6311c51dd8916a5d122045de02b6cc8e8e7684"} Feb 25 13:37:16 crc kubenswrapper[4815]: I0225 13:37:16.961895 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-b9qrv" podStartSLOduration=2.732337671 podStartE2EDuration="16.961878663s" podCreationTimestamp="2026-02-25 13:37:00 +0000 UTC" firstStartedPulling="2026-02-25 13:37:01.814046319 +0000 UTC m=+979.615144373" lastFinishedPulling="2026-02-25 13:37:16.043587301 +0000 UTC m=+993.844685365" observedRunningTime="2026-02-25 13:37:16.959410996 +0000 UTC m=+994.760509050" watchObservedRunningTime="2026-02-25 13:37:16.961878663 +0000 UTC m=+994.762976717" Feb 25 13:37:17 crc kubenswrapper[4815]: I0225 13:37:17.015317 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371964.839481 podStartE2EDuration="1m12.015295394s" podCreationTimestamp="2026-02-25 13:36:05 +0000 UTC" firstStartedPulling="2026-02-25 13:36:07.704641161 +0000 UTC m=+925.505739215" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:37:17.009143963 +0000 UTC m=+994.810242017" watchObservedRunningTime="2026-02-25 13:37:17.015295394 +0000 UTC m=+994.816393448" Feb 25 13:37:17 crc kubenswrapper[4815]: I0225 13:37:17.927645 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" event={"ID":"05c3c613-64ee-4c4d-a7b0-16265fc7ba09","Type":"ContainerStarted","Data":"b51d2b4b4b9c20b230c81658fa2798ac35458848948144bd83fc04590a2bef0f"} Feb 25 13:37:17 crc kubenswrapper[4815]: I0225 13:37:17.960731 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" podStartSLOduration=9.960664643 podStartE2EDuration="9.960664643s" podCreationTimestamp="2026-02-25 13:37:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:37:17.953235273 +0000 UTC m=+995.754333337" watchObservedRunningTime="2026-02-25 13:37:17.960664643 +0000 UTC m=+995.761762697" Feb 25 13:37:18 crc kubenswrapper[4815]: I0225 13:37:18.334455 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-49c8k" Feb 25 13:37:18 crc kubenswrapper[4815]: I0225 13:37:18.477740 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wnld\" (UniqueName: \"kubernetes.io/projected/6dbd3d76-08d8-4798-91ab-2dc5afe1993f-kube-api-access-7wnld\") pod \"6dbd3d76-08d8-4798-91ab-2dc5afe1993f\" (UID: \"6dbd3d76-08d8-4798-91ab-2dc5afe1993f\") " Feb 25 13:37:18 crc kubenswrapper[4815]: I0225 13:37:18.477980 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dbd3d76-08d8-4798-91ab-2dc5afe1993f-operator-scripts\") pod \"6dbd3d76-08d8-4798-91ab-2dc5afe1993f\" (UID: \"6dbd3d76-08d8-4798-91ab-2dc5afe1993f\") " Feb 25 13:37:18 crc kubenswrapper[4815]: I0225 13:37:18.478399 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6dbd3d76-08d8-4798-91ab-2dc5afe1993f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6dbd3d76-08d8-4798-91ab-2dc5afe1993f" (UID: "6dbd3d76-08d8-4798-91ab-2dc5afe1993f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:18 crc kubenswrapper[4815]: I0225 13:37:18.498314 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dbd3d76-08d8-4798-91ab-2dc5afe1993f-kube-api-access-7wnld" (OuterVolumeSpecName: "kube-api-access-7wnld") pod "6dbd3d76-08d8-4798-91ab-2dc5afe1993f" (UID: "6dbd3d76-08d8-4798-91ab-2dc5afe1993f"). InnerVolumeSpecName "kube-api-access-7wnld". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:37:18 crc kubenswrapper[4815]: I0225 13:37:18.512283 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:18 crc kubenswrapper[4815]: I0225 13:37:18.580399 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wnld\" (UniqueName: \"kubernetes.io/projected/6dbd3d76-08d8-4798-91ab-2dc5afe1993f-kube-api-access-7wnld\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:18 crc kubenswrapper[4815]: I0225 13:37:18.580440 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dbd3d76-08d8-4798-91ab-2dc5afe1993f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:18 crc kubenswrapper[4815]: I0225 13:37:18.952306 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-49c8k" Feb 25 13:37:18 crc kubenswrapper[4815]: I0225 13:37:18.952305 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-49c8k" event={"ID":"6dbd3d76-08d8-4798-91ab-2dc5afe1993f","Type":"ContainerDied","Data":"9a3e4c6876284fd8aa7436efcc41991076c3d959e60643e420cfbe42d1a44ad5"} Feb 25 13:37:18 crc kubenswrapper[4815]: I0225 13:37:18.952720 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a3e4c6876284fd8aa7436efcc41991076c3d959e60643e420cfbe42d1a44ad5" Feb 25 13:37:21 crc kubenswrapper[4815]: I0225 13:37:21.342745 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:37:21 crc kubenswrapper[4815]: I0225 13:37:21.342830 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:37:21 crc kubenswrapper[4815]: I0225 13:37:21.342917 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:37:21 crc kubenswrapper[4815]: I0225 13:37:21.343896 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"65cc59f637c0af5e4fc8efb7df93ca963c3aa3c1788d5699757ce0a0287dce17"} pod="openshift-machine-config-operator/machine-config-daemon-frc75" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 13:37:21 crc kubenswrapper[4815]: I0225 13:37:21.344005 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" containerID="cri-o://65cc59f637c0af5e4fc8efb7df93ca963c3aa3c1788d5699757ce0a0287dce17" gracePeriod=600 Feb 25 13:37:21 crc kubenswrapper[4815]: I0225 13:37:21.990978 4815 generic.go:334] "Generic (PLEG): container finished" podID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerID="65cc59f637c0af5e4fc8efb7df93ca963c3aa3c1788d5699757ce0a0287dce17" exitCode=0 Feb 25 13:37:21 crc kubenswrapper[4815]: I0225 13:37:21.991055 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerDied","Data":"65cc59f637c0af5e4fc8efb7df93ca963c3aa3c1788d5699757ce0a0287dce17"} Feb 25 13:37:21 crc kubenswrapper[4815]: I0225 13:37:21.991356 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerStarted","Data":"90119f271263ed85c7a4bae34f95ffae1fb69e1997807ded5781ce1fee85b5cc"} Feb 25 13:37:21 crc kubenswrapper[4815]: I0225 13:37:21.991421 4815 scope.go:117] "RemoveContainer" containerID="6557b63a0f423077dc6748b76d4eac34119edc35e845268afc85f7e306b0281a" Feb 25 13:37:23 crc kubenswrapper[4815]: I0225 13:37:23.008833 4815 generic.go:334] "Generic (PLEG): container finished" podID="7e9844cb-1b51-4e42-8fee-5ed7059776c1" containerID="f2018e423eff7617abcb82018b6311c51dd8916a5d122045de02b6cc8e8e7684" exitCode=0 Feb 25 13:37:23 crc kubenswrapper[4815]: I0225 13:37:23.008959 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-b9qrv" event={"ID":"7e9844cb-1b51-4e42-8fee-5ed7059776c1","Type":"ContainerDied","Data":"f2018e423eff7617abcb82018b6311c51dd8916a5d122045de02b6cc8e8e7684"} Feb 25 13:37:23 crc kubenswrapper[4815]: I0225 13:37:23.513942 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:23 crc kubenswrapper[4815]: I0225 13:37:23.588542 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-kj6n4"] Feb 25 13:37:23 crc kubenswrapper[4815]: I0225 13:37:23.588805 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" podUID="09095c00-a8b0-4c47-84e5-4db2eb1d4309" containerName="dnsmasq-dns" containerID="cri-o://cd5bddfd310cfee75c571904678ee904d02b05cac2419967fc5d5a8ca4c24d27" gracePeriod=10 Feb 25 13:37:23 crc kubenswrapper[4815]: I0225 13:37:23.985693 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.020577 4815 generic.go:334] "Generic (PLEG): container finished" podID="09095c00-a8b0-4c47-84e5-4db2eb1d4309" containerID="cd5bddfd310cfee75c571904678ee904d02b05cac2419967fc5d5a8ca4c24d27" exitCode=0 Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.020880 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.021669 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" event={"ID":"09095c00-a8b0-4c47-84e5-4db2eb1d4309","Type":"ContainerDied","Data":"cd5bddfd310cfee75c571904678ee904d02b05cac2419967fc5d5a8ca4c24d27"} Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.021718 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-kj6n4" event={"ID":"09095c00-a8b0-4c47-84e5-4db2eb1d4309","Type":"ContainerDied","Data":"41ba47c4105756c83948b222e4d7bbfd1ce535823a8c281931c3c84cbf3bd4a1"} Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.021752 4815 scope.go:117] "RemoveContainer" containerID="cd5bddfd310cfee75c571904678ee904d02b05cac2419967fc5d5a8ca4c24d27" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.063330 4815 scope.go:117] "RemoveContainer" containerID="7422c646f8f65fd0c5a074ce28df635475cd197cb888b59eeaf7567c25aac492" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.081301 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-ovsdbserver-sb\") pod \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.081368 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2824z\" (UniqueName: \"kubernetes.io/projected/09095c00-a8b0-4c47-84e5-4db2eb1d4309-kube-api-access-2824z\") pod \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.081436 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-ovsdbserver-nb\") pod \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.081638 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-config\") pod \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.081725 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-dns-svc\") pod \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\" (UID: \"09095c00-a8b0-4c47-84e5-4db2eb1d4309\") " Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.086882 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09095c00-a8b0-4c47-84e5-4db2eb1d4309-kube-api-access-2824z" (OuterVolumeSpecName: "kube-api-access-2824z") pod "09095c00-a8b0-4c47-84e5-4db2eb1d4309" (UID: "09095c00-a8b0-4c47-84e5-4db2eb1d4309"). InnerVolumeSpecName "kube-api-access-2824z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.087433 4815 scope.go:117] "RemoveContainer" containerID="cd5bddfd310cfee75c571904678ee904d02b05cac2419967fc5d5a8ca4c24d27" Feb 25 13:37:24 crc kubenswrapper[4815]: E0225 13:37:24.087966 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd5bddfd310cfee75c571904678ee904d02b05cac2419967fc5d5a8ca4c24d27\": container with ID starting with cd5bddfd310cfee75c571904678ee904d02b05cac2419967fc5d5a8ca4c24d27 not found: ID does not exist" containerID="cd5bddfd310cfee75c571904678ee904d02b05cac2419967fc5d5a8ca4c24d27" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.088086 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd5bddfd310cfee75c571904678ee904d02b05cac2419967fc5d5a8ca4c24d27"} err="failed to get container status \"cd5bddfd310cfee75c571904678ee904d02b05cac2419967fc5d5a8ca4c24d27\": rpc error: code = NotFound desc = could not find container \"cd5bddfd310cfee75c571904678ee904d02b05cac2419967fc5d5a8ca4c24d27\": container with ID starting with cd5bddfd310cfee75c571904678ee904d02b05cac2419967fc5d5a8ca4c24d27 not found: ID does not exist" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.088194 4815 scope.go:117] "RemoveContainer" containerID="7422c646f8f65fd0c5a074ce28df635475cd197cb888b59eeaf7567c25aac492" Feb 25 13:37:24 crc kubenswrapper[4815]: E0225 13:37:24.088942 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7422c646f8f65fd0c5a074ce28df635475cd197cb888b59eeaf7567c25aac492\": container with ID starting with 7422c646f8f65fd0c5a074ce28df635475cd197cb888b59eeaf7567c25aac492 not found: ID does not exist" containerID="7422c646f8f65fd0c5a074ce28df635475cd197cb888b59eeaf7567c25aac492" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.089026 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7422c646f8f65fd0c5a074ce28df635475cd197cb888b59eeaf7567c25aac492"} err="failed to get container status \"7422c646f8f65fd0c5a074ce28df635475cd197cb888b59eeaf7567c25aac492\": rpc error: code = NotFound desc = could not find container \"7422c646f8f65fd0c5a074ce28df635475cd197cb888b59eeaf7567c25aac492\": container with ID starting with 7422c646f8f65fd0c5a074ce28df635475cd197cb888b59eeaf7567c25aac492 not found: ID does not exist" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.131085 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-config" (OuterVolumeSpecName: "config") pod "09095c00-a8b0-4c47-84e5-4db2eb1d4309" (UID: "09095c00-a8b0-4c47-84e5-4db2eb1d4309"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.146101 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "09095c00-a8b0-4c47-84e5-4db2eb1d4309" (UID: "09095c00-a8b0-4c47-84e5-4db2eb1d4309"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.146605 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "09095c00-a8b0-4c47-84e5-4db2eb1d4309" (UID: "09095c00-a8b0-4c47-84e5-4db2eb1d4309"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.148433 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "09095c00-a8b0-4c47-84e5-4db2eb1d4309" (UID: "09095c00-a8b0-4c47-84e5-4db2eb1d4309"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.183802 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.183840 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.183852 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.183866 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2824z\" (UniqueName: \"kubernetes.io/projected/09095c00-a8b0-4c47-84e5-4db2eb1d4309-kube-api-access-2824z\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.183879 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/09095c00-a8b0-4c47-84e5-4db2eb1d4309-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.357031 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-kj6n4"] Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.367994 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-kj6n4"] Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.383770 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-b9qrv" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.487797 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-db-sync-config-data\") pod \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\" (UID: \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\") " Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.488116 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxvhb\" (UniqueName: \"kubernetes.io/projected/7e9844cb-1b51-4e42-8fee-5ed7059776c1-kube-api-access-cxvhb\") pod \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\" (UID: \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\") " Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.488701 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-combined-ca-bundle\") pod \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\" (UID: \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\") " Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.488839 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-config-data\") pod \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\" (UID: \"7e9844cb-1b51-4e42-8fee-5ed7059776c1\") " Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.492018 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e9844cb-1b51-4e42-8fee-5ed7059776c1-kube-api-access-cxvhb" (OuterVolumeSpecName: "kube-api-access-cxvhb") pod "7e9844cb-1b51-4e42-8fee-5ed7059776c1" (UID: "7e9844cb-1b51-4e42-8fee-5ed7059776c1"). InnerVolumeSpecName "kube-api-access-cxvhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.492231 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7e9844cb-1b51-4e42-8fee-5ed7059776c1" (UID: "7e9844cb-1b51-4e42-8fee-5ed7059776c1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.511678 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e9844cb-1b51-4e42-8fee-5ed7059776c1" (UID: "7e9844cb-1b51-4e42-8fee-5ed7059776c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.532710 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-config-data" (OuterVolumeSpecName: "config-data") pod "7e9844cb-1b51-4e42-8fee-5ed7059776c1" (UID: "7e9844cb-1b51-4e42-8fee-5ed7059776c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.590568 4815 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.590841 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxvhb\" (UniqueName: \"kubernetes.io/projected/7e9844cb-1b51-4e42-8fee-5ed7059776c1-kube-api-access-cxvhb\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.590913 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.590976 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e9844cb-1b51-4e42-8fee-5ed7059776c1-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:24 crc kubenswrapper[4815]: I0225 13:37:24.948452 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09095c00-a8b0-4c47-84e5-4db2eb1d4309" path="/var/lib/kubelet/pods/09095c00-a8b0-4c47-84e5-4db2eb1d4309/volumes" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.038345 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-b9qrv" event={"ID":"7e9844cb-1b51-4e42-8fee-5ed7059776c1","Type":"ContainerDied","Data":"de6a14523ff7330a0720239d4ab5eb7b7d52576f29f265da5a7d99b2090f91c3"} Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.038389 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de6a14523ff7330a0720239d4ab5eb7b7d52576f29f265da5a7d99b2090f91c3" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.038495 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-b9qrv" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.487678 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-5lnz6"] Feb 25 13:37:25 crc kubenswrapper[4815]: E0225 13:37:25.488210 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9844cb-1b51-4e42-8fee-5ed7059776c1" containerName="glance-db-sync" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.488225 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9844cb-1b51-4e42-8fee-5ed7059776c1" containerName="glance-db-sync" Feb 25 13:37:25 crc kubenswrapper[4815]: E0225 13:37:25.488254 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09095c00-a8b0-4c47-84e5-4db2eb1d4309" containerName="init" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.488260 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="09095c00-a8b0-4c47-84e5-4db2eb1d4309" containerName="init" Feb 25 13:37:25 crc kubenswrapper[4815]: E0225 13:37:25.488268 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dbd3d76-08d8-4798-91ab-2dc5afe1993f" containerName="mariadb-account-create-update" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.488276 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dbd3d76-08d8-4798-91ab-2dc5afe1993f" containerName="mariadb-account-create-update" Feb 25 13:37:25 crc kubenswrapper[4815]: E0225 13:37:25.488288 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09095c00-a8b0-4c47-84e5-4db2eb1d4309" containerName="dnsmasq-dns" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.488294 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="09095c00-a8b0-4c47-84e5-4db2eb1d4309" containerName="dnsmasq-dns" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.488427 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e9844cb-1b51-4e42-8fee-5ed7059776c1" containerName="glance-db-sync" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.488441 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dbd3d76-08d8-4798-91ab-2dc5afe1993f" containerName="mariadb-account-create-update" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.488456 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="09095c00-a8b0-4c47-84e5-4db2eb1d4309" containerName="dnsmasq-dns" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.489191 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.516464 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-5lnz6"] Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.608319 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-dns-svc\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.608376 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngwbf\" (UniqueName: \"kubernetes.io/projected/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-kube-api-access-ngwbf\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.608393 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-config\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.608435 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-dns-swift-storage-0\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.608471 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-ovsdbserver-sb\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.608526 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-ovsdbserver-nb\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.709609 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-ovsdbserver-sb\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.709692 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-ovsdbserver-nb\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.709749 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-dns-svc\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.709776 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngwbf\" (UniqueName: \"kubernetes.io/projected/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-kube-api-access-ngwbf\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.709792 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-config\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.709833 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-dns-swift-storage-0\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.711095 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-ovsdbserver-sb\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.711691 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-ovsdbserver-nb\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.711754 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-dns-svc\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.711802 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-dns-swift-storage-0\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.712214 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-config\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.751673 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngwbf\" (UniqueName: \"kubernetes.io/projected/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-kube-api-access-ngwbf\") pod \"dnsmasq-dns-895cf5cf-5lnz6\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:25 crc kubenswrapper[4815]: I0225 13:37:25.805754 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:26 crc kubenswrapper[4815]: I0225 13:37:26.278670 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-5lnz6"] Feb 25 13:37:26 crc kubenswrapper[4815]: W0225 13:37:26.288037 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3199ac65_2fd7_482f_b2fa_b5305ffa59d8.slice/crio-a0aa0fb524582fe979521956cdf24a03caa513bc0d1459cb333dc11d2ea37f61 WatchSource:0}: Error finding container a0aa0fb524582fe979521956cdf24a03caa513bc0d1459cb333dc11d2ea37f61: Status 404 returned error can't find the container with id a0aa0fb524582fe979521956cdf24a03caa513bc0d1459cb333dc11d2ea37f61 Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.057686 4815 generic.go:334] "Generic (PLEG): container finished" podID="3199ac65-2fd7-482f-b2fa-b5305ffa59d8" containerID="4706cd5360f951abc9d3faa9a0e981097dc091f1e28467b705fee874282528dc" exitCode=0 Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.057834 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" event={"ID":"3199ac65-2fd7-482f-b2fa-b5305ffa59d8","Type":"ContainerDied","Data":"4706cd5360f951abc9d3faa9a0e981097dc091f1e28467b705fee874282528dc"} Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.058169 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" event={"ID":"3199ac65-2fd7-482f-b2fa-b5305ffa59d8","Type":"ContainerStarted","Data":"a0aa0fb524582fe979521956cdf24a03caa513bc0d1459cb333dc11d2ea37f61"} Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.338729 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.638602 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-bc7lz"] Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.639594 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bc7lz" Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.649186 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-bc7lz"] Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.750663 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvlpc\" (UniqueName: \"kubernetes.io/projected/d7b99305-c70e-422b-925c-d9abfc038a5e-kube-api-access-pvlpc\") pod \"cinder-db-create-bc7lz\" (UID: \"d7b99305-c70e-422b-925c-d9abfc038a5e\") " pod="openstack/cinder-db-create-bc7lz" Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.750923 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b99305-c70e-422b-925c-d9abfc038a5e-operator-scripts\") pod \"cinder-db-create-bc7lz\" (UID: \"d7b99305-c70e-422b-925c-d9abfc038a5e\") " pod="openstack/cinder-db-create-bc7lz" Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.853216 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvlpc\" (UniqueName: \"kubernetes.io/projected/d7b99305-c70e-422b-925c-d9abfc038a5e-kube-api-access-pvlpc\") pod \"cinder-db-create-bc7lz\" (UID: \"d7b99305-c70e-422b-925c-d9abfc038a5e\") " pod="openstack/cinder-db-create-bc7lz" Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.853320 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b99305-c70e-422b-925c-d9abfc038a5e-operator-scripts\") pod \"cinder-db-create-bc7lz\" (UID: \"d7b99305-c70e-422b-925c-d9abfc038a5e\") " pod="openstack/cinder-db-create-bc7lz" Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.854201 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b99305-c70e-422b-925c-d9abfc038a5e-operator-scripts\") pod \"cinder-db-create-bc7lz\" (UID: \"d7b99305-c70e-422b-925c-d9abfc038a5e\") " pod="openstack/cinder-db-create-bc7lz" Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.870993 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-55a3-account-create-update-vjcsr"] Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.871928 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-55a3-account-create-update-vjcsr" Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.879303 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvlpc\" (UniqueName: \"kubernetes.io/projected/d7b99305-c70e-422b-925c-d9abfc038a5e-kube-api-access-pvlpc\") pod \"cinder-db-create-bc7lz\" (UID: \"d7b99305-c70e-422b-925c-d9abfc038a5e\") " pod="openstack/cinder-db-create-bc7lz" Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.892547 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.894130 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-55a3-account-create-update-vjcsr"] Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.955565 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjsvx\" (UniqueName: \"kubernetes.io/projected/80bad2a7-eed6-4d5e-a631-160681e10e41-kube-api-access-wjsvx\") pod \"cinder-55a3-account-create-update-vjcsr\" (UID: \"80bad2a7-eed6-4d5e-a631-160681e10e41\") " pod="openstack/cinder-55a3-account-create-update-vjcsr" Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.955647 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80bad2a7-eed6-4d5e-a631-160681e10e41-operator-scripts\") pod \"cinder-55a3-account-create-update-vjcsr\" (UID: \"80bad2a7-eed6-4d5e-a631-160681e10e41\") " pod="openstack/cinder-55a3-account-create-update-vjcsr" Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.956119 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bc7lz" Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.957499 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-6hvck"] Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.958687 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-6hvck" Feb 25 13:37:27 crc kubenswrapper[4815]: I0225 13:37:27.967799 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-6hvck"] Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.050997 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-lrkj8"] Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.051991 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lrkj8" Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.057809 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a45cad8-f93f-4bf1-9fc8-d382e6be15f7-operator-scripts\") pod \"barbican-db-create-6hvck\" (UID: \"7a45cad8-f93f-4bf1-9fc8-d382e6be15f7\") " pod="openstack/barbican-db-create-6hvck" Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.058134 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xplz\" (UniqueName: \"kubernetes.io/projected/7a45cad8-f93f-4bf1-9fc8-d382e6be15f7-kube-api-access-4xplz\") pod \"barbican-db-create-6hvck\" (UID: \"7a45cad8-f93f-4bf1-9fc8-d382e6be15f7\") " pod="openstack/barbican-db-create-6hvck" Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.058250 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjsvx\" (UniqueName: \"kubernetes.io/projected/80bad2a7-eed6-4d5e-a631-160681e10e41-kube-api-access-wjsvx\") pod \"cinder-55a3-account-create-update-vjcsr\" (UID: \"80bad2a7-eed6-4d5e-a631-160681e10e41\") " pod="openstack/cinder-55a3-account-create-update-vjcsr" Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.058343 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80bad2a7-eed6-4d5e-a631-160681e10e41-operator-scripts\") pod \"cinder-55a3-account-create-update-vjcsr\" (UID: \"80bad2a7-eed6-4d5e-a631-160681e10e41\") " pod="openstack/cinder-55a3-account-create-update-vjcsr" Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.059470 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80bad2a7-eed6-4d5e-a631-160681e10e41-operator-scripts\") pod \"cinder-55a3-account-create-update-vjcsr\" (UID: \"80bad2a7-eed6-4d5e-a631-160681e10e41\") " pod="openstack/cinder-55a3-account-create-update-vjcsr" Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.076217 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-lrkj8"] Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.082607 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-szm5v"] Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.083819 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-szm5v" Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.089784 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.089963 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.090086 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.090650 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-tftcl" Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.100594 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-198e-account-create-update-t9r4s"] Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.101707 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-198e-account-create-update-t9r4s" Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.109364 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-198e-account-create-update-t9r4s"] Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.113834 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.114331 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" event={"ID":"3199ac65-2fd7-482f-b2fa-b5305ffa59d8","Type":"ContainerStarted","Data":"35b893c825ea2cf30ddd270904b104304eed3db2ef55461fe4ab5b3edeb5c804"} Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.118172 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-szm5v"] Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.127289 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjsvx\" (UniqueName: \"kubernetes.io/projected/80bad2a7-eed6-4d5e-a631-160681e10e41-kube-api-access-wjsvx\") pod \"cinder-55a3-account-create-update-vjcsr\" (UID: \"80bad2a7-eed6-4d5e-a631-160681e10e41\") " pod="openstack/cinder-55a3-account-create-update-vjcsr" Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.200551 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xplz\" (UniqueName: \"kubernetes.io/projected/7a45cad8-f93f-4bf1-9fc8-d382e6be15f7-kube-api-access-4xplz\") pod \"barbican-db-create-6hvck\" (UID: \"7a45cad8-f93f-4bf1-9fc8-d382e6be15f7\") " pod="openstack/barbican-db-create-6hvck" Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.200936 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a45cad8-f93f-4bf1-9fc8-d382e6be15f7-operator-scripts\") pod \"barbican-db-create-6hvck\" (UID: \"7a45cad8-f93f-4bf1-9fc8-d382e6be15f7\") " pod="openstack/barbican-db-create-6hvck" Feb 25 13:37:28 crc kubenswrapper[4815]: I0225 13:37:28.201700 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a45cad8-f93f-4bf1-9fc8-d382e6be15f7-operator-scripts\") pod \"barbican-db-create-6hvck\" (UID: \"7a45cad8-f93f-4bf1-9fc8-d382e6be15f7\") " pod="openstack/barbican-db-create-6hvck" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.251789 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-ce70-account-create-update-75n7s"] Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.252434 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xplz\" (UniqueName: \"kubernetes.io/projected/7a45cad8-f93f-4bf1-9fc8-d382e6be15f7-kube-api-access-4xplz\") pod \"barbican-db-create-6hvck\" (UID: \"7a45cad8-f93f-4bf1-9fc8-d382e6be15f7\") " pod="openstack/barbican-db-create-6hvck" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.253091 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ce70-account-create-update-75n7s" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.254296 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-55a3-account-create-update-vjcsr" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.261279 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-ce70-account-create-update-75n7s"] Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.262808 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.278811 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-6hvck" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.302485 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b912f0ba-4a07-4b2a-b372-a107335c4b73-operator-scripts\") pod \"neutron-198e-account-create-update-t9r4s\" (UID: \"b912f0ba-4a07-4b2a-b372-a107335c4b73\") " pod="openstack/neutron-198e-account-create-update-t9r4s" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.302552 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e394112-2dbe-43b0-8d71-5763831f591a-operator-scripts\") pod \"barbican-ce70-account-create-update-75n7s\" (UID: \"2e394112-2dbe-43b0-8d71-5763831f591a\") " pod="openstack/barbican-ce70-account-create-update-75n7s" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.302577 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rn2c\" (UniqueName: \"kubernetes.io/projected/b912f0ba-4a07-4b2a-b372-a107335c4b73-kube-api-access-2rn2c\") pod \"neutron-198e-account-create-update-t9r4s\" (UID: \"b912f0ba-4a07-4b2a-b372-a107335c4b73\") " pod="openstack/neutron-198e-account-create-update-t9r4s" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.302601 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjw6n\" (UniqueName: \"kubernetes.io/projected/75268cc3-204e-4ac5-9cbd-958343dd2011-kube-api-access-jjw6n\") pod \"keystone-db-sync-szm5v\" (UID: \"75268cc3-204e-4ac5-9cbd-958343dd2011\") " pod="openstack/keystone-db-sync-szm5v" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.302629 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75268cc3-204e-4ac5-9cbd-958343dd2011-combined-ca-bundle\") pod \"keystone-db-sync-szm5v\" (UID: \"75268cc3-204e-4ac5-9cbd-958343dd2011\") " pod="openstack/keystone-db-sync-szm5v" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.302648 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxpkz\" (UniqueName: \"kubernetes.io/projected/1095ac90-8cd8-41d8-8eb7-847f72d3d82b-kube-api-access-pxpkz\") pod \"neutron-db-create-lrkj8\" (UID: \"1095ac90-8cd8-41d8-8eb7-847f72d3d82b\") " pod="openstack/neutron-db-create-lrkj8" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.302690 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1095ac90-8cd8-41d8-8eb7-847f72d3d82b-operator-scripts\") pod \"neutron-db-create-lrkj8\" (UID: \"1095ac90-8cd8-41d8-8eb7-847f72d3d82b\") " pod="openstack/neutron-db-create-lrkj8" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.302716 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xq5d\" (UniqueName: \"kubernetes.io/projected/2e394112-2dbe-43b0-8d71-5763831f591a-kube-api-access-9xq5d\") pod \"barbican-ce70-account-create-update-75n7s\" (UID: \"2e394112-2dbe-43b0-8d71-5763831f591a\") " pod="openstack/barbican-ce70-account-create-update-75n7s" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.302751 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75268cc3-204e-4ac5-9cbd-958343dd2011-config-data\") pod \"keystone-db-sync-szm5v\" (UID: \"75268cc3-204e-4ac5-9cbd-958343dd2011\") " pod="openstack/keystone-db-sync-szm5v" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.404315 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b912f0ba-4a07-4b2a-b372-a107335c4b73-operator-scripts\") pod \"neutron-198e-account-create-update-t9r4s\" (UID: \"b912f0ba-4a07-4b2a-b372-a107335c4b73\") " pod="openstack/neutron-198e-account-create-update-t9r4s" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.404356 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e394112-2dbe-43b0-8d71-5763831f591a-operator-scripts\") pod \"barbican-ce70-account-create-update-75n7s\" (UID: \"2e394112-2dbe-43b0-8d71-5763831f591a\") " pod="openstack/barbican-ce70-account-create-update-75n7s" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.404371 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rn2c\" (UniqueName: \"kubernetes.io/projected/b912f0ba-4a07-4b2a-b372-a107335c4b73-kube-api-access-2rn2c\") pod \"neutron-198e-account-create-update-t9r4s\" (UID: \"b912f0ba-4a07-4b2a-b372-a107335c4b73\") " pod="openstack/neutron-198e-account-create-update-t9r4s" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.404389 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjw6n\" (UniqueName: \"kubernetes.io/projected/75268cc3-204e-4ac5-9cbd-958343dd2011-kube-api-access-jjw6n\") pod \"keystone-db-sync-szm5v\" (UID: \"75268cc3-204e-4ac5-9cbd-958343dd2011\") " pod="openstack/keystone-db-sync-szm5v" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.404411 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75268cc3-204e-4ac5-9cbd-958343dd2011-combined-ca-bundle\") pod \"keystone-db-sync-szm5v\" (UID: \"75268cc3-204e-4ac5-9cbd-958343dd2011\") " pod="openstack/keystone-db-sync-szm5v" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.404426 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxpkz\" (UniqueName: \"kubernetes.io/projected/1095ac90-8cd8-41d8-8eb7-847f72d3d82b-kube-api-access-pxpkz\") pod \"neutron-db-create-lrkj8\" (UID: \"1095ac90-8cd8-41d8-8eb7-847f72d3d82b\") " pod="openstack/neutron-db-create-lrkj8" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.404457 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1095ac90-8cd8-41d8-8eb7-847f72d3d82b-operator-scripts\") pod \"neutron-db-create-lrkj8\" (UID: \"1095ac90-8cd8-41d8-8eb7-847f72d3d82b\") " pod="openstack/neutron-db-create-lrkj8" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.404477 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xq5d\" (UniqueName: \"kubernetes.io/projected/2e394112-2dbe-43b0-8d71-5763831f591a-kube-api-access-9xq5d\") pod \"barbican-ce70-account-create-update-75n7s\" (UID: \"2e394112-2dbe-43b0-8d71-5763831f591a\") " pod="openstack/barbican-ce70-account-create-update-75n7s" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.404514 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75268cc3-204e-4ac5-9cbd-958343dd2011-config-data\") pod \"keystone-db-sync-szm5v\" (UID: \"75268cc3-204e-4ac5-9cbd-958343dd2011\") " pod="openstack/keystone-db-sync-szm5v" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.406700 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e394112-2dbe-43b0-8d71-5763831f591a-operator-scripts\") pod \"barbican-ce70-account-create-update-75n7s\" (UID: \"2e394112-2dbe-43b0-8d71-5763831f591a\") " pod="openstack/barbican-ce70-account-create-update-75n7s" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.407003 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1095ac90-8cd8-41d8-8eb7-847f72d3d82b-operator-scripts\") pod \"neutron-db-create-lrkj8\" (UID: \"1095ac90-8cd8-41d8-8eb7-847f72d3d82b\") " pod="openstack/neutron-db-create-lrkj8" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.407253 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b912f0ba-4a07-4b2a-b372-a107335c4b73-operator-scripts\") pod \"neutron-198e-account-create-update-t9r4s\" (UID: \"b912f0ba-4a07-4b2a-b372-a107335c4b73\") " pod="openstack/neutron-198e-account-create-update-t9r4s" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.411224 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75268cc3-204e-4ac5-9cbd-958343dd2011-config-data\") pod \"keystone-db-sync-szm5v\" (UID: \"75268cc3-204e-4ac5-9cbd-958343dd2011\") " pod="openstack/keystone-db-sync-szm5v" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.413288 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75268cc3-204e-4ac5-9cbd-958343dd2011-combined-ca-bundle\") pod \"keystone-db-sync-szm5v\" (UID: \"75268cc3-204e-4ac5-9cbd-958343dd2011\") " pod="openstack/keystone-db-sync-szm5v" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.425793 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rn2c\" (UniqueName: \"kubernetes.io/projected/b912f0ba-4a07-4b2a-b372-a107335c4b73-kube-api-access-2rn2c\") pod \"neutron-198e-account-create-update-t9r4s\" (UID: \"b912f0ba-4a07-4b2a-b372-a107335c4b73\") " pod="openstack/neutron-198e-account-create-update-t9r4s" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.431152 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxpkz\" (UniqueName: \"kubernetes.io/projected/1095ac90-8cd8-41d8-8eb7-847f72d3d82b-kube-api-access-pxpkz\") pod \"neutron-db-create-lrkj8\" (UID: \"1095ac90-8cd8-41d8-8eb7-847f72d3d82b\") " pod="openstack/neutron-db-create-lrkj8" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.444112 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjw6n\" (UniqueName: \"kubernetes.io/projected/75268cc3-204e-4ac5-9cbd-958343dd2011-kube-api-access-jjw6n\") pod \"keystone-db-sync-szm5v\" (UID: \"75268cc3-204e-4ac5-9cbd-958343dd2011\") " pod="openstack/keystone-db-sync-szm5v" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.445112 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xq5d\" (UniqueName: \"kubernetes.io/projected/2e394112-2dbe-43b0-8d71-5763831f591a-kube-api-access-9xq5d\") pod \"barbican-ce70-account-create-update-75n7s\" (UID: \"2e394112-2dbe-43b0-8d71-5763831f591a\") " pod="openstack/barbican-ce70-account-create-update-75n7s" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.468554 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-szm5v" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.472033 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-bc7lz"] Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.503313 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-198e-account-create-update-t9r4s" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.610356 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ce70-account-create-update-75n7s" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:28.674906 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lrkj8" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:29.122383 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bc7lz" event={"ID":"d7b99305-c70e-422b-925c-d9abfc038a5e","Type":"ContainerStarted","Data":"8f2acee70c2106cfd7c5d91b5b6fafe3a1ecd9b8a9f71cb5b9fa2362a78d8f1e"} Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:31.163180 4815 generic.go:334] "Generic (PLEG): container finished" podID="d7b99305-c70e-422b-925c-d9abfc038a5e" containerID="a6084341b178e86ba2f4e30ef013ac0618b5b575ca28c75ef03b15cd6a863aaa" exitCode=0 Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:31.163273 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bc7lz" event={"ID":"d7b99305-c70e-422b-925c-d9abfc038a5e","Type":"ContainerDied","Data":"a6084341b178e86ba2f4e30ef013ac0618b5b575ca28c75ef03b15cd6a863aaa"} Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:31.163869 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:31.214488 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" podStartSLOduration=6.214457356 podStartE2EDuration="6.214457356s" podCreationTimestamp="2026-02-25 13:37:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:37:31.209294736 +0000 UTC m=+1009.010392830" watchObservedRunningTime="2026-02-25 13:37:31.214457356 +0000 UTC m=+1009.015555450" Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:31.806298 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-55a3-account-create-update-vjcsr"] Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:31.826541 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-lrkj8"] Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:31.857601 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-6hvck"] Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:31.871912 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-198e-account-create-update-t9r4s"] Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:31.881110 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-ce70-account-create-update-75n7s"] Feb 25 13:37:31 crc kubenswrapper[4815]: I0225 13:37:31.887160 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-szm5v"] Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.172036 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-198e-account-create-update-t9r4s" event={"ID":"b912f0ba-4a07-4b2a-b372-a107335c4b73","Type":"ContainerStarted","Data":"16f6265b307311743431a947d38dd872750903a01059624c86677b8b18aa1581"} Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.172355 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-198e-account-create-update-t9r4s" event={"ID":"b912f0ba-4a07-4b2a-b372-a107335c4b73","Type":"ContainerStarted","Data":"fe3f7bfe8767b5f108da1b9a62209a193657d6dc3bfeb0fae27b265123a42a19"} Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.175214 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lrkj8" event={"ID":"1095ac90-8cd8-41d8-8eb7-847f72d3d82b","Type":"ContainerStarted","Data":"9e48c2504573e6f135e3796d7ae6705ac4f3d7b6a34fe030c1ba321ee632844a"} Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.175263 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lrkj8" event={"ID":"1095ac90-8cd8-41d8-8eb7-847f72d3d82b","Type":"ContainerStarted","Data":"03c23db8e48769becb328af414b0877479d9348eeef2c3fe897a3bfe22c9332f"} Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.179102 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-55a3-account-create-update-vjcsr" event={"ID":"80bad2a7-eed6-4d5e-a631-160681e10e41","Type":"ContainerStarted","Data":"412bba9239a0b0e8d3a6b7639584f77e2da023f720ddb4cad3e9fbbde36761b0"} Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.179156 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-55a3-account-create-update-vjcsr" event={"ID":"80bad2a7-eed6-4d5e-a631-160681e10e41","Type":"ContainerStarted","Data":"e327bac14fbbcf890848dfae184bf32272bad0da1f03db393a7e82f8be27c2c6"} Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.180865 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-szm5v" event={"ID":"75268cc3-204e-4ac5-9cbd-958343dd2011","Type":"ContainerStarted","Data":"d31eab3dd3b060fc66a6e800ec72f70e738781d0068cbd459427165da3746ca3"} Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.182536 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ce70-account-create-update-75n7s" event={"ID":"2e394112-2dbe-43b0-8d71-5763831f591a","Type":"ContainerStarted","Data":"e7700b21016181b8e1d769c3aee588c4178c7221decd73961bca470edef59d89"} Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.182560 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ce70-account-create-update-75n7s" event={"ID":"2e394112-2dbe-43b0-8d71-5763831f591a","Type":"ContainerStarted","Data":"fe547ac5470e4c1c47e4e3b2d30d76c90d52fc81a949e075df2e817425b0baeb"} Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.184959 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-6hvck" event={"ID":"7a45cad8-f93f-4bf1-9fc8-d382e6be15f7","Type":"ContainerStarted","Data":"af0e9370809232daebcb9ce84280e56c4c962c934ca4372cff8f06aadf0c8259"} Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.185017 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-6hvck" event={"ID":"7a45cad8-f93f-4bf1-9fc8-d382e6be15f7","Type":"ContainerStarted","Data":"0582aaadb4e038725247ffb802978f16c7ed83bb71d26f64fb62caf845fdaa60"} Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.185626 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.196439 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-198e-account-create-update-t9r4s" podStartSLOduration=4.196417206 podStartE2EDuration="4.196417206s" podCreationTimestamp="2026-02-25 13:37:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:37:32.195081934 +0000 UTC m=+1009.996179988" watchObservedRunningTime="2026-02-25 13:37:32.196417206 +0000 UTC m=+1009.997515260" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.244577 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-ce70-account-create-update-75n7s" podStartSLOduration=4.244554563 podStartE2EDuration="4.244554563s" podCreationTimestamp="2026-02-25 13:37:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:37:32.239400604 +0000 UTC m=+1010.040498658" watchObservedRunningTime="2026-02-25 13:37:32.244554563 +0000 UTC m=+1010.045652617" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.255887 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-swzkm"] Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.256103 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" podUID="05c3c613-64ee-4c4d-a7b0-16265fc7ba09" containerName="dnsmasq-dns" containerID="cri-o://b51d2b4b4b9c20b230c81658fa2798ac35458848948144bd83fc04590a2bef0f" gracePeriod=10 Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.263343 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-6hvck" podStartSLOduration=5.263328474 podStartE2EDuration="5.263328474s" podCreationTimestamp="2026-02-25 13:37:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:37:32.259838616 +0000 UTC m=+1010.060936670" watchObservedRunningTime="2026-02-25 13:37:32.263328474 +0000 UTC m=+1010.064426528" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.295094 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-lrkj8" podStartSLOduration=4.295077544 podStartE2EDuration="4.295077544s" podCreationTimestamp="2026-02-25 13:37:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:37:32.290826373 +0000 UTC m=+1010.091924427" watchObservedRunningTime="2026-02-25 13:37:32.295077544 +0000 UTC m=+1010.096175598" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.324161 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-55a3-account-create-update-vjcsr" podStartSLOduration=5.324140363 podStartE2EDuration="5.324140363s" podCreationTimestamp="2026-02-25 13:37:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:37:32.31593842 +0000 UTC m=+1010.117036484" watchObservedRunningTime="2026-02-25 13:37:32.324140363 +0000 UTC m=+1010.125238417" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.684856 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bc7lz" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.691156 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.791615 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-config\") pod \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.791723 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-dns-svc\") pod \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.791743 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-ovsdbserver-sb\") pod \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.791784 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b99305-c70e-422b-925c-d9abfc038a5e-operator-scripts\") pod \"d7b99305-c70e-422b-925c-d9abfc038a5e\" (UID: \"d7b99305-c70e-422b-925c-d9abfc038a5e\") " Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.791834 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5x5kh\" (UniqueName: \"kubernetes.io/projected/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-kube-api-access-5x5kh\") pod \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.791863 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-ovsdbserver-nb\") pod \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.791925 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-dns-swift-storage-0\") pod \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\" (UID: \"05c3c613-64ee-4c4d-a7b0-16265fc7ba09\") " Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.791963 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvlpc\" (UniqueName: \"kubernetes.io/projected/d7b99305-c70e-422b-925c-d9abfc038a5e-kube-api-access-pvlpc\") pod \"d7b99305-c70e-422b-925c-d9abfc038a5e\" (UID: \"d7b99305-c70e-422b-925c-d9abfc038a5e\") " Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.792278 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7b99305-c70e-422b-925c-d9abfc038a5e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d7b99305-c70e-422b-925c-d9abfc038a5e" (UID: "d7b99305-c70e-422b-925c-d9abfc038a5e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.797877 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-kube-api-access-5x5kh" (OuterVolumeSpecName: "kube-api-access-5x5kh") pod "05c3c613-64ee-4c4d-a7b0-16265fc7ba09" (UID: "05c3c613-64ee-4c4d-a7b0-16265fc7ba09"). InnerVolumeSpecName "kube-api-access-5x5kh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.800212 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7b99305-c70e-422b-925c-d9abfc038a5e-kube-api-access-pvlpc" (OuterVolumeSpecName: "kube-api-access-pvlpc") pod "d7b99305-c70e-422b-925c-d9abfc038a5e" (UID: "d7b99305-c70e-422b-925c-d9abfc038a5e"). InnerVolumeSpecName "kube-api-access-pvlpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.838157 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "05c3c613-64ee-4c4d-a7b0-16265fc7ba09" (UID: "05c3c613-64ee-4c4d-a7b0-16265fc7ba09"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.842384 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "05c3c613-64ee-4c4d-a7b0-16265fc7ba09" (UID: "05c3c613-64ee-4c4d-a7b0-16265fc7ba09"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.849055 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-config" (OuterVolumeSpecName: "config") pod "05c3c613-64ee-4c4d-a7b0-16265fc7ba09" (UID: "05c3c613-64ee-4c4d-a7b0-16265fc7ba09"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.850055 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "05c3c613-64ee-4c4d-a7b0-16265fc7ba09" (UID: "05c3c613-64ee-4c4d-a7b0-16265fc7ba09"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.854867 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "05c3c613-64ee-4c4d-a7b0-16265fc7ba09" (UID: "05c3c613-64ee-4c4d-a7b0-16265fc7ba09"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.894394 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5x5kh\" (UniqueName: \"kubernetes.io/projected/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-kube-api-access-5x5kh\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.894419 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.894430 4815 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.894439 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvlpc\" (UniqueName: \"kubernetes.io/projected/d7b99305-c70e-422b-925c-d9abfc038a5e-kube-api-access-pvlpc\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.894448 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.894455 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.894463 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05c3c613-64ee-4c4d-a7b0-16265fc7ba09-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:32 crc kubenswrapper[4815]: I0225 13:37:32.894471 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b99305-c70e-422b-925c-d9abfc038a5e-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.195832 4815 generic.go:334] "Generic (PLEG): container finished" podID="7a45cad8-f93f-4bf1-9fc8-d382e6be15f7" containerID="af0e9370809232daebcb9ce84280e56c4c962c934ca4372cff8f06aadf0c8259" exitCode=0 Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.195887 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-6hvck" event={"ID":"7a45cad8-f93f-4bf1-9fc8-d382e6be15f7","Type":"ContainerDied","Data":"af0e9370809232daebcb9ce84280e56c4c962c934ca4372cff8f06aadf0c8259"} Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.197671 4815 generic.go:334] "Generic (PLEG): container finished" podID="b912f0ba-4a07-4b2a-b372-a107335c4b73" containerID="16f6265b307311743431a947d38dd872750903a01059624c86677b8b18aa1581" exitCode=0 Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.197830 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-198e-account-create-update-t9r4s" event={"ID":"b912f0ba-4a07-4b2a-b372-a107335c4b73","Type":"ContainerDied","Data":"16f6265b307311743431a947d38dd872750903a01059624c86677b8b18aa1581"} Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.199805 4815 generic.go:334] "Generic (PLEG): container finished" podID="1095ac90-8cd8-41d8-8eb7-847f72d3d82b" containerID="9e48c2504573e6f135e3796d7ae6705ac4f3d7b6a34fe030c1ba321ee632844a" exitCode=0 Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.199845 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lrkj8" event={"ID":"1095ac90-8cd8-41d8-8eb7-847f72d3d82b","Type":"ContainerDied","Data":"9e48c2504573e6f135e3796d7ae6705ac4f3d7b6a34fe030c1ba321ee632844a"} Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.201412 4815 generic.go:334] "Generic (PLEG): container finished" podID="80bad2a7-eed6-4d5e-a631-160681e10e41" containerID="412bba9239a0b0e8d3a6b7639584f77e2da023f720ddb4cad3e9fbbde36761b0" exitCode=0 Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.201560 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-55a3-account-create-update-vjcsr" event={"ID":"80bad2a7-eed6-4d5e-a631-160681e10e41","Type":"ContainerDied","Data":"412bba9239a0b0e8d3a6b7639584f77e2da023f720ddb4cad3e9fbbde36761b0"} Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.204354 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bc7lz" event={"ID":"d7b99305-c70e-422b-925c-d9abfc038a5e","Type":"ContainerDied","Data":"8f2acee70c2106cfd7c5d91b5b6fafe3a1ecd9b8a9f71cb5b9fa2362a78d8f1e"} Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.204375 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f2acee70c2106cfd7c5d91b5b6fafe3a1ecd9b8a9f71cb5b9fa2362a78d8f1e" Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.204400 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bc7lz" Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.206356 4815 generic.go:334] "Generic (PLEG): container finished" podID="05c3c613-64ee-4c4d-a7b0-16265fc7ba09" containerID="b51d2b4b4b9c20b230c81658fa2798ac35458848948144bd83fc04590a2bef0f" exitCode=0 Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.206397 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" event={"ID":"05c3c613-64ee-4c4d-a7b0-16265fc7ba09","Type":"ContainerDied","Data":"b51d2b4b4b9c20b230c81658fa2798ac35458848948144bd83fc04590a2bef0f"} Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.206424 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" event={"ID":"05c3c613-64ee-4c4d-a7b0-16265fc7ba09","Type":"ContainerDied","Data":"ce044b46fef43b20c233823110d650e5a6b5c48e7207609ed2dc874167a3735a"} Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.206442 4815 scope.go:117] "RemoveContainer" containerID="b51d2b4b4b9c20b230c81658fa2798ac35458848948144bd83fc04590a2bef0f" Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.206438 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-swzkm" Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.208188 4815 generic.go:334] "Generic (PLEG): container finished" podID="2e394112-2dbe-43b0-8d71-5763831f591a" containerID="e7700b21016181b8e1d769c3aee588c4178c7221decd73961bca470edef59d89" exitCode=0 Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.208902 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ce70-account-create-update-75n7s" event={"ID":"2e394112-2dbe-43b0-8d71-5763831f591a","Type":"ContainerDied","Data":"e7700b21016181b8e1d769c3aee588c4178c7221decd73961bca470edef59d89"} Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.228186 4815 scope.go:117] "RemoveContainer" containerID="4ba253a99bac5cd3e4f35dc846673196ba5ef58e76e79061dabd4217083cee4d" Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.244484 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-swzkm"] Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.259684 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-swzkm"] Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.298229 4815 scope.go:117] "RemoveContainer" containerID="b51d2b4b4b9c20b230c81658fa2798ac35458848948144bd83fc04590a2bef0f" Feb 25 13:37:33 crc kubenswrapper[4815]: E0225 13:37:33.301632 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b51d2b4b4b9c20b230c81658fa2798ac35458848948144bd83fc04590a2bef0f\": container with ID starting with b51d2b4b4b9c20b230c81658fa2798ac35458848948144bd83fc04590a2bef0f not found: ID does not exist" containerID="b51d2b4b4b9c20b230c81658fa2798ac35458848948144bd83fc04590a2bef0f" Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.301714 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b51d2b4b4b9c20b230c81658fa2798ac35458848948144bd83fc04590a2bef0f"} err="failed to get container status \"b51d2b4b4b9c20b230c81658fa2798ac35458848948144bd83fc04590a2bef0f\": rpc error: code = NotFound desc = could not find container \"b51d2b4b4b9c20b230c81658fa2798ac35458848948144bd83fc04590a2bef0f\": container with ID starting with b51d2b4b4b9c20b230c81658fa2798ac35458848948144bd83fc04590a2bef0f not found: ID does not exist" Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.301745 4815 scope.go:117] "RemoveContainer" containerID="4ba253a99bac5cd3e4f35dc846673196ba5ef58e76e79061dabd4217083cee4d" Feb 25 13:37:33 crc kubenswrapper[4815]: E0225 13:37:33.302529 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ba253a99bac5cd3e4f35dc846673196ba5ef58e76e79061dabd4217083cee4d\": container with ID starting with 4ba253a99bac5cd3e4f35dc846673196ba5ef58e76e79061dabd4217083cee4d not found: ID does not exist" containerID="4ba253a99bac5cd3e4f35dc846673196ba5ef58e76e79061dabd4217083cee4d" Feb 25 13:37:33 crc kubenswrapper[4815]: I0225 13:37:33.302632 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ba253a99bac5cd3e4f35dc846673196ba5ef58e76e79061dabd4217083cee4d"} err="failed to get container status \"4ba253a99bac5cd3e4f35dc846673196ba5ef58e76e79061dabd4217083cee4d\": rpc error: code = NotFound desc = could not find container \"4ba253a99bac5cd3e4f35dc846673196ba5ef58e76e79061dabd4217083cee4d\": container with ID starting with 4ba253a99bac5cd3e4f35dc846673196ba5ef58e76e79061dabd4217083cee4d not found: ID does not exist" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.603849 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-55a3-account-create-update-vjcsr" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.737684 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80bad2a7-eed6-4d5e-a631-160681e10e41-operator-scripts\") pod \"80bad2a7-eed6-4d5e-a631-160681e10e41\" (UID: \"80bad2a7-eed6-4d5e-a631-160681e10e41\") " Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.737792 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjsvx\" (UniqueName: \"kubernetes.io/projected/80bad2a7-eed6-4d5e-a631-160681e10e41-kube-api-access-wjsvx\") pod \"80bad2a7-eed6-4d5e-a631-160681e10e41\" (UID: \"80bad2a7-eed6-4d5e-a631-160681e10e41\") " Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.740181 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80bad2a7-eed6-4d5e-a631-160681e10e41-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "80bad2a7-eed6-4d5e-a631-160681e10e41" (UID: "80bad2a7-eed6-4d5e-a631-160681e10e41"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.743700 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80bad2a7-eed6-4d5e-a631-160681e10e41-kube-api-access-wjsvx" (OuterVolumeSpecName: "kube-api-access-wjsvx") pod "80bad2a7-eed6-4d5e-a631-160681e10e41" (UID: "80bad2a7-eed6-4d5e-a631-160681e10e41"). InnerVolumeSpecName "kube-api-access-wjsvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.815145 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-6hvck" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.820906 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ce70-account-create-update-75n7s" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.827161 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lrkj8" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.839921 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80bad2a7-eed6-4d5e-a631-160681e10e41-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.839945 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjsvx\" (UniqueName: \"kubernetes.io/projected/80bad2a7-eed6-4d5e-a631-160681e10e41-kube-api-access-wjsvx\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.845437 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-198e-account-create-update-t9r4s" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.944519 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1095ac90-8cd8-41d8-8eb7-847f72d3d82b-operator-scripts\") pod \"1095ac90-8cd8-41d8-8eb7-847f72d3d82b\" (UID: \"1095ac90-8cd8-41d8-8eb7-847f72d3d82b\") " Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.944575 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rn2c\" (UniqueName: \"kubernetes.io/projected/b912f0ba-4a07-4b2a-b372-a107335c4b73-kube-api-access-2rn2c\") pod \"b912f0ba-4a07-4b2a-b372-a107335c4b73\" (UID: \"b912f0ba-4a07-4b2a-b372-a107335c4b73\") " Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.944696 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e394112-2dbe-43b0-8d71-5763831f591a-operator-scripts\") pod \"2e394112-2dbe-43b0-8d71-5763831f591a\" (UID: \"2e394112-2dbe-43b0-8d71-5763831f591a\") " Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.944723 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxpkz\" (UniqueName: \"kubernetes.io/projected/1095ac90-8cd8-41d8-8eb7-847f72d3d82b-kube-api-access-pxpkz\") pod \"1095ac90-8cd8-41d8-8eb7-847f72d3d82b\" (UID: \"1095ac90-8cd8-41d8-8eb7-847f72d3d82b\") " Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.944739 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a45cad8-f93f-4bf1-9fc8-d382e6be15f7-operator-scripts\") pod \"7a45cad8-f93f-4bf1-9fc8-d382e6be15f7\" (UID: \"7a45cad8-f93f-4bf1-9fc8-d382e6be15f7\") " Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.944781 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xq5d\" (UniqueName: \"kubernetes.io/projected/2e394112-2dbe-43b0-8d71-5763831f591a-kube-api-access-9xq5d\") pod \"2e394112-2dbe-43b0-8d71-5763831f591a\" (UID: \"2e394112-2dbe-43b0-8d71-5763831f591a\") " Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.944798 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xplz\" (UniqueName: \"kubernetes.io/projected/7a45cad8-f93f-4bf1-9fc8-d382e6be15f7-kube-api-access-4xplz\") pod \"7a45cad8-f93f-4bf1-9fc8-d382e6be15f7\" (UID: \"7a45cad8-f93f-4bf1-9fc8-d382e6be15f7\") " Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.944856 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b912f0ba-4a07-4b2a-b372-a107335c4b73-operator-scripts\") pod \"b912f0ba-4a07-4b2a-b372-a107335c4b73\" (UID: \"b912f0ba-4a07-4b2a-b372-a107335c4b73\") " Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.944895 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1095ac90-8cd8-41d8-8eb7-847f72d3d82b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1095ac90-8cd8-41d8-8eb7-847f72d3d82b" (UID: "1095ac90-8cd8-41d8-8eb7-847f72d3d82b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.945155 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e394112-2dbe-43b0-8d71-5763831f591a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2e394112-2dbe-43b0-8d71-5763831f591a" (UID: "2e394112-2dbe-43b0-8d71-5763831f591a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.945277 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05c3c613-64ee-4c4d-a7b0-16265fc7ba09" path="/var/lib/kubelet/pods/05c3c613-64ee-4c4d-a7b0-16265fc7ba09/volumes" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.945361 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b912f0ba-4a07-4b2a-b372-a107335c4b73-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b912f0ba-4a07-4b2a-b372-a107335c4b73" (UID: "b912f0ba-4a07-4b2a-b372-a107335c4b73"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.945533 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b912f0ba-4a07-4b2a-b372-a107335c4b73-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.945555 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1095ac90-8cd8-41d8-8eb7-847f72d3d82b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.945565 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e394112-2dbe-43b0-8d71-5763831f591a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.945878 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a45cad8-f93f-4bf1-9fc8-d382e6be15f7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7a45cad8-f93f-4bf1-9fc8-d382e6be15f7" (UID: "7a45cad8-f93f-4bf1-9fc8-d382e6be15f7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.948245 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b912f0ba-4a07-4b2a-b372-a107335c4b73-kube-api-access-2rn2c" (OuterVolumeSpecName: "kube-api-access-2rn2c") pod "b912f0ba-4a07-4b2a-b372-a107335c4b73" (UID: "b912f0ba-4a07-4b2a-b372-a107335c4b73"). InnerVolumeSpecName "kube-api-access-2rn2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.948289 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e394112-2dbe-43b0-8d71-5763831f591a-kube-api-access-9xq5d" (OuterVolumeSpecName: "kube-api-access-9xq5d") pod "2e394112-2dbe-43b0-8d71-5763831f591a" (UID: "2e394112-2dbe-43b0-8d71-5763831f591a"). InnerVolumeSpecName "kube-api-access-9xq5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.948343 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a45cad8-f93f-4bf1-9fc8-d382e6be15f7-kube-api-access-4xplz" (OuterVolumeSpecName: "kube-api-access-4xplz") pod "7a45cad8-f93f-4bf1-9fc8-d382e6be15f7" (UID: "7a45cad8-f93f-4bf1-9fc8-d382e6be15f7"). InnerVolumeSpecName "kube-api-access-4xplz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:37:34 crc kubenswrapper[4815]: I0225 13:37:34.948993 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1095ac90-8cd8-41d8-8eb7-847f72d3d82b-kube-api-access-pxpkz" (OuterVolumeSpecName: "kube-api-access-pxpkz") pod "1095ac90-8cd8-41d8-8eb7-847f72d3d82b" (UID: "1095ac90-8cd8-41d8-8eb7-847f72d3d82b"). InnerVolumeSpecName "kube-api-access-pxpkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.047584 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rn2c\" (UniqueName: \"kubernetes.io/projected/b912f0ba-4a07-4b2a-b372-a107335c4b73-kube-api-access-2rn2c\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.047616 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxpkz\" (UniqueName: \"kubernetes.io/projected/1095ac90-8cd8-41d8-8eb7-847f72d3d82b-kube-api-access-pxpkz\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.047626 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a45cad8-f93f-4bf1-9fc8-d382e6be15f7-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.047635 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xq5d\" (UniqueName: \"kubernetes.io/projected/2e394112-2dbe-43b0-8d71-5763831f591a-kube-api-access-9xq5d\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.047643 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xplz\" (UniqueName: \"kubernetes.io/projected/7a45cad8-f93f-4bf1-9fc8-d382e6be15f7-kube-api-access-4xplz\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.232835 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-198e-account-create-update-t9r4s" event={"ID":"b912f0ba-4a07-4b2a-b372-a107335c4b73","Type":"ContainerDied","Data":"fe3f7bfe8767b5f108da1b9a62209a193657d6dc3bfeb0fae27b265123a42a19"} Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.232876 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe3f7bfe8767b5f108da1b9a62209a193657d6dc3bfeb0fae27b265123a42a19" Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.232874 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-198e-account-create-update-t9r4s" Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.234942 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-lrkj8" event={"ID":"1095ac90-8cd8-41d8-8eb7-847f72d3d82b","Type":"ContainerDied","Data":"03c23db8e48769becb328af414b0877479d9348eeef2c3fe897a3bfe22c9332f"} Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.234977 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03c23db8e48769becb328af414b0877479d9348eeef2c3fe897a3bfe22c9332f" Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.234981 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-lrkj8" Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.236981 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-55a3-account-create-update-vjcsr" event={"ID":"80bad2a7-eed6-4d5e-a631-160681e10e41","Type":"ContainerDied","Data":"e327bac14fbbcf890848dfae184bf32272bad0da1f03db393a7e82f8be27c2c6"} Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.237006 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e327bac14fbbcf890848dfae184bf32272bad0da1f03db393a7e82f8be27c2c6" Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.237053 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-55a3-account-create-update-vjcsr" Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.241326 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ce70-account-create-update-75n7s" event={"ID":"2e394112-2dbe-43b0-8d71-5763831f591a","Type":"ContainerDied","Data":"fe547ac5470e4c1c47e4e3b2d30d76c90d52fc81a949e075df2e817425b0baeb"} Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.241347 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe547ac5470e4c1c47e4e3b2d30d76c90d52fc81a949e075df2e817425b0baeb" Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.241386 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ce70-account-create-update-75n7s" Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.247291 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-6hvck" event={"ID":"7a45cad8-f93f-4bf1-9fc8-d382e6be15f7","Type":"ContainerDied","Data":"0582aaadb4e038725247ffb802978f16c7ed83bb71d26f64fb62caf845fdaa60"} Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.247340 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0582aaadb4e038725247ffb802978f16c7ed83bb71d26f64fb62caf845fdaa60" Feb 25 13:37:35 crc kubenswrapper[4815]: I0225 13:37:35.247383 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-6hvck" Feb 25 13:37:38 crc kubenswrapper[4815]: I0225 13:37:38.300878 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-szm5v" event={"ID":"75268cc3-204e-4ac5-9cbd-958343dd2011","Type":"ContainerStarted","Data":"4fe814b94077353d6a8d2f48e8ce07000a7d18da25c3e7efec37bce1ec6f0c06"} Feb 25 13:37:38 crc kubenswrapper[4815]: I0225 13:37:38.320570 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-szm5v" podStartSLOduration=4.261108675 podStartE2EDuration="10.320551333s" podCreationTimestamp="2026-02-25 13:37:28 +0000 UTC" firstStartedPulling="2026-02-25 13:37:31.860454762 +0000 UTC m=+1009.661552816" lastFinishedPulling="2026-02-25 13:37:37.91989738 +0000 UTC m=+1015.720995474" observedRunningTime="2026-02-25 13:37:38.316730935 +0000 UTC m=+1016.117828989" watchObservedRunningTime="2026-02-25 13:37:38.320551333 +0000 UTC m=+1016.121649387" Feb 25 13:37:41 crc kubenswrapper[4815]: I0225 13:37:41.337598 4815 generic.go:334] "Generic (PLEG): container finished" podID="75268cc3-204e-4ac5-9cbd-958343dd2011" containerID="4fe814b94077353d6a8d2f48e8ce07000a7d18da25c3e7efec37bce1ec6f0c06" exitCode=0 Feb 25 13:37:41 crc kubenswrapper[4815]: I0225 13:37:41.337839 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-szm5v" event={"ID":"75268cc3-204e-4ac5-9cbd-958343dd2011","Type":"ContainerDied","Data":"4fe814b94077353d6a8d2f48e8ce07000a7d18da25c3e7efec37bce1ec6f0c06"} Feb 25 13:37:42 crc kubenswrapper[4815]: I0225 13:37:42.795737 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-szm5v" Feb 25 13:37:42 crc kubenswrapper[4815]: I0225 13:37:42.906158 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjw6n\" (UniqueName: \"kubernetes.io/projected/75268cc3-204e-4ac5-9cbd-958343dd2011-kube-api-access-jjw6n\") pod \"75268cc3-204e-4ac5-9cbd-958343dd2011\" (UID: \"75268cc3-204e-4ac5-9cbd-958343dd2011\") " Feb 25 13:37:42 crc kubenswrapper[4815]: I0225 13:37:42.906227 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75268cc3-204e-4ac5-9cbd-958343dd2011-combined-ca-bundle\") pod \"75268cc3-204e-4ac5-9cbd-958343dd2011\" (UID: \"75268cc3-204e-4ac5-9cbd-958343dd2011\") " Feb 25 13:37:42 crc kubenswrapper[4815]: I0225 13:37:42.906321 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75268cc3-204e-4ac5-9cbd-958343dd2011-config-data\") pod \"75268cc3-204e-4ac5-9cbd-958343dd2011\" (UID: \"75268cc3-204e-4ac5-9cbd-958343dd2011\") " Feb 25 13:37:42 crc kubenswrapper[4815]: I0225 13:37:42.913868 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75268cc3-204e-4ac5-9cbd-958343dd2011-kube-api-access-jjw6n" (OuterVolumeSpecName: "kube-api-access-jjw6n") pod "75268cc3-204e-4ac5-9cbd-958343dd2011" (UID: "75268cc3-204e-4ac5-9cbd-958343dd2011"). InnerVolumeSpecName "kube-api-access-jjw6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:37:42 crc kubenswrapper[4815]: I0225 13:37:42.938192 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75268cc3-204e-4ac5-9cbd-958343dd2011-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75268cc3-204e-4ac5-9cbd-958343dd2011" (UID: "75268cc3-204e-4ac5-9cbd-958343dd2011"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:37:42 crc kubenswrapper[4815]: I0225 13:37:42.980404 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75268cc3-204e-4ac5-9cbd-958343dd2011-config-data" (OuterVolumeSpecName: "config-data") pod "75268cc3-204e-4ac5-9cbd-958343dd2011" (UID: "75268cc3-204e-4ac5-9cbd-958343dd2011"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.007872 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjw6n\" (UniqueName: \"kubernetes.io/projected/75268cc3-204e-4ac5-9cbd-958343dd2011-kube-api-access-jjw6n\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.007909 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75268cc3-204e-4ac5-9cbd-958343dd2011-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.007918 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75268cc3-204e-4ac5-9cbd-958343dd2011-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.360663 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-szm5v" event={"ID":"75268cc3-204e-4ac5-9cbd-958343dd2011","Type":"ContainerDied","Data":"d31eab3dd3b060fc66a6e800ec72f70e738781d0068cbd459427165da3746ca3"} Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.360712 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-szm5v" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.360728 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d31eab3dd3b060fc66a6e800ec72f70e738781d0068cbd459427165da3746ca3" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.747974 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-chzbc"] Feb 25 13:37:43 crc kubenswrapper[4815]: E0225 13:37:43.749085 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05c3c613-64ee-4c4d-a7b0-16265fc7ba09" containerName="init" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.749268 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="05c3c613-64ee-4c4d-a7b0-16265fc7ba09" containerName="init" Feb 25 13:37:43 crc kubenswrapper[4815]: E0225 13:37:43.749335 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a45cad8-f93f-4bf1-9fc8-d382e6be15f7" containerName="mariadb-database-create" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.749396 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a45cad8-f93f-4bf1-9fc8-d382e6be15f7" containerName="mariadb-database-create" Feb 25 13:37:43 crc kubenswrapper[4815]: E0225 13:37:43.749474 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b912f0ba-4a07-4b2a-b372-a107335c4b73" containerName="mariadb-account-create-update" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.749553 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b912f0ba-4a07-4b2a-b372-a107335c4b73" containerName="mariadb-account-create-update" Feb 25 13:37:43 crc kubenswrapper[4815]: E0225 13:37:43.749630 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7b99305-c70e-422b-925c-d9abfc038a5e" containerName="mariadb-database-create" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.749714 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7b99305-c70e-422b-925c-d9abfc038a5e" containerName="mariadb-database-create" Feb 25 13:37:43 crc kubenswrapper[4815]: E0225 13:37:43.749800 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75268cc3-204e-4ac5-9cbd-958343dd2011" containerName="keystone-db-sync" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.749864 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="75268cc3-204e-4ac5-9cbd-958343dd2011" containerName="keystone-db-sync" Feb 25 13:37:43 crc kubenswrapper[4815]: E0225 13:37:43.749936 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1095ac90-8cd8-41d8-8eb7-847f72d3d82b" containerName="mariadb-database-create" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.749997 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1095ac90-8cd8-41d8-8eb7-847f72d3d82b" containerName="mariadb-database-create" Feb 25 13:37:43 crc kubenswrapper[4815]: E0225 13:37:43.750986 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e394112-2dbe-43b0-8d71-5763831f591a" containerName="mariadb-account-create-update" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.751071 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e394112-2dbe-43b0-8d71-5763831f591a" containerName="mariadb-account-create-update" Feb 25 13:37:43 crc kubenswrapper[4815]: E0225 13:37:43.751140 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80bad2a7-eed6-4d5e-a631-160681e10e41" containerName="mariadb-account-create-update" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.751194 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="80bad2a7-eed6-4d5e-a631-160681e10e41" containerName="mariadb-account-create-update" Feb 25 13:37:43 crc kubenswrapper[4815]: E0225 13:37:43.751262 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05c3c613-64ee-4c4d-a7b0-16265fc7ba09" containerName="dnsmasq-dns" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.751326 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="05c3c613-64ee-4c4d-a7b0-16265fc7ba09" containerName="dnsmasq-dns" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.751663 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b912f0ba-4a07-4b2a-b372-a107335c4b73" containerName="mariadb-account-create-update" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.751745 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="80bad2a7-eed6-4d5e-a631-160681e10e41" containerName="mariadb-account-create-update" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.751800 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7b99305-c70e-422b-925c-d9abfc038a5e" containerName="mariadb-database-create" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.751856 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a45cad8-f93f-4bf1-9fc8-d382e6be15f7" containerName="mariadb-database-create" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.751920 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1095ac90-8cd8-41d8-8eb7-847f72d3d82b" containerName="mariadb-database-create" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.751979 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="75268cc3-204e-4ac5-9cbd-958343dd2011" containerName="keystone-db-sync" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.752042 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e394112-2dbe-43b0-8d71-5763831f591a" containerName="mariadb-account-create-update" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.752195 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="05c3c613-64ee-4c4d-a7b0-16265fc7ba09" containerName="dnsmasq-dns" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.754412 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.774302 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-chzbc"] Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.811662 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-qs8h9"] Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.812613 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.816000 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.816244 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.818294 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.818424 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-tftcl" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.822366 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.832847 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-qs8h9"] Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.842560 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.842760 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw2v6\" (UniqueName: \"kubernetes.io/projected/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-kube-api-access-rw2v6\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.842825 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-combined-ca-bundle\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.842921 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-fernet-keys\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.843015 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-scripts\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.843095 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-dns-swift-storage-0\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.843169 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xkm4\" (UniqueName: \"kubernetes.io/projected/e8df64e8-8db3-4a60-807d-f8cf502320e4-kube-api-access-2xkm4\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.843237 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-config\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.843298 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-dns-svc\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.843355 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-config-data\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.843431 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-ovsdbserver-nb\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.843529 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-credential-keys\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.946933 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-scripts\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.946988 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-dns-swift-storage-0\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.947013 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xkm4\" (UniqueName: \"kubernetes.io/projected/e8df64e8-8db3-4a60-807d-f8cf502320e4-kube-api-access-2xkm4\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.947038 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-dns-svc\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.947053 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-config\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.947068 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-config-data\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.947102 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-ovsdbserver-nb\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.947132 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-credential-keys\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.947158 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.947179 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw2v6\" (UniqueName: \"kubernetes.io/projected/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-kube-api-access-rw2v6\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.947192 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-combined-ca-bundle\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.947209 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-fernet-keys\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.954393 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-fernet-keys\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.956357 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.956900 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-ovsdbserver-nb\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.961583 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-scripts\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.962540 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-dns-swift-storage-0\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.963082 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-dns-svc\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.963586 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-config\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.964634 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-combined-ca-bundle\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.964645 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-config-data\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.965266 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-credential-keys\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.969050 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-2kr57"] Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.969980 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2kr57" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.989854 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.990035 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 25 13:37:43 crc kubenswrapper[4815]: I0225 13:37:43.990178 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qmrd2" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.007826 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xkm4\" (UniqueName: \"kubernetes.io/projected/e8df64e8-8db3-4a60-807d-f8cf502320e4-kube-api-access-2xkm4\") pod \"dnsmasq-dns-6c9c9f998c-chzbc\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.024569 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-c455f8ff5-tkdz5"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.025953 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.030887 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.031131 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-zswd4" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.031241 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.031355 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.037911 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw2v6\" (UniqueName: \"kubernetes.io/projected/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-kube-api-access-rw2v6\") pod \"keystone-bootstrap-qs8h9\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.040055 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2kr57"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.073423 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c455f8ff5-tkdz5"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.082915 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.115232 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.117200 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.123888 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.124116 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.140690 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.152905 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d9c18809-9f62-4b7e-8103-df382e9b4fa6-horizon-secret-key\") pod \"horizon-c455f8ff5-tkdz5\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.153111 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab03b49b-be69-4e1d-99e7-2f71a602b110-combined-ca-bundle\") pod \"neutron-db-sync-2kr57\" (UID: \"ab03b49b-be69-4e1d-99e7-2f71a602b110\") " pod="openstack/neutron-db-sync-2kr57" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.153232 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9c18809-9f62-4b7e-8103-df382e9b4fa6-logs\") pod \"horizon-c455f8ff5-tkdz5\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.153310 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d9c18809-9f62-4b7e-8103-df382e9b4fa6-config-data\") pod \"horizon-c455f8ff5-tkdz5\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.153389 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d9c18809-9f62-4b7e-8103-df382e9b4fa6-scripts\") pod \"horizon-c455f8ff5-tkdz5\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.153468 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ab03b49b-be69-4e1d-99e7-2f71a602b110-config\") pod \"neutron-db-sync-2kr57\" (UID: \"ab03b49b-be69-4e1d-99e7-2f71a602b110\") " pod="openstack/neutron-db-sync-2kr57" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.153715 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmp45\" (UniqueName: \"kubernetes.io/projected/d9c18809-9f62-4b7e-8103-df382e9b4fa6-kube-api-access-cmp45\") pod \"horizon-c455f8ff5-tkdz5\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.153805 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v8w7\" (UniqueName: \"kubernetes.io/projected/ab03b49b-be69-4e1d-99e7-2f71a602b110-kube-api-access-9v8w7\") pod \"neutron-db-sync-2kr57\" (UID: \"ab03b49b-be69-4e1d-99e7-2f71a602b110\") " pod="openstack/neutron-db-sync-2kr57" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.157896 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.200860 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-576gs"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.202291 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.211640 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-zjkm2" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.211853 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.211964 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.249305 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-chzbc"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.255444 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v8w7\" (UniqueName: \"kubernetes.io/projected/ab03b49b-be69-4e1d-99e7-2f71a602b110-kube-api-access-9v8w7\") pod \"neutron-db-sync-2kr57\" (UID: \"ab03b49b-be69-4e1d-99e7-2f71a602b110\") " pod="openstack/neutron-db-sync-2kr57" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.255485 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3643568-6595-4c73-9c83-be6e28fabe83-log-httpd\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.255520 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3643568-6595-4c73-9c83-be6e28fabe83-run-httpd\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.255572 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mv9lz\" (UniqueName: \"kubernetes.io/projected/f3643568-6595-4c73-9c83-be6e28fabe83-kube-api-access-mv9lz\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.255593 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab03b49b-be69-4e1d-99e7-2f71a602b110-combined-ca-bundle\") pod \"neutron-db-sync-2kr57\" (UID: \"ab03b49b-be69-4e1d-99e7-2f71a602b110\") " pod="openstack/neutron-db-sync-2kr57" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.255610 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d9c18809-9f62-4b7e-8103-df382e9b4fa6-horizon-secret-key\") pod \"horizon-c455f8ff5-tkdz5\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.255635 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.255674 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9c18809-9f62-4b7e-8103-df382e9b4fa6-logs\") pod \"horizon-c455f8ff5-tkdz5\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.255697 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d9c18809-9f62-4b7e-8103-df382e9b4fa6-config-data\") pod \"horizon-c455f8ff5-tkdz5\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.255721 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d9c18809-9f62-4b7e-8103-df382e9b4fa6-scripts\") pod \"horizon-c455f8ff5-tkdz5\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.255737 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.255764 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ab03b49b-be69-4e1d-99e7-2f71a602b110-config\") pod \"neutron-db-sync-2kr57\" (UID: \"ab03b49b-be69-4e1d-99e7-2f71a602b110\") " pod="openstack/neutron-db-sync-2kr57" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.255793 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-scripts\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.255821 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-config-data\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.255847 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmp45\" (UniqueName: \"kubernetes.io/projected/d9c18809-9f62-4b7e-8103-df382e9b4fa6-kube-api-access-cmp45\") pod \"horizon-c455f8ff5-tkdz5\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.258874 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d9c18809-9f62-4b7e-8103-df382e9b4fa6-scripts\") pod \"horizon-c455f8ff5-tkdz5\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.259093 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9c18809-9f62-4b7e-8103-df382e9b4fa6-logs\") pod \"horizon-c455f8ff5-tkdz5\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.259963 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d9c18809-9f62-4b7e-8103-df382e9b4fa6-config-data\") pod \"horizon-c455f8ff5-tkdz5\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.282076 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-576gs"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.290015 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ab03b49b-be69-4e1d-99e7-2f71a602b110-config\") pod \"neutron-db-sync-2kr57\" (UID: \"ab03b49b-be69-4e1d-99e7-2f71a602b110\") " pod="openstack/neutron-db-sync-2kr57" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.290561 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d9c18809-9f62-4b7e-8103-df382e9b4fa6-horizon-secret-key\") pod \"horizon-c455f8ff5-tkdz5\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.291109 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab03b49b-be69-4e1d-99e7-2f71a602b110-combined-ca-bundle\") pod \"neutron-db-sync-2kr57\" (UID: \"ab03b49b-be69-4e1d-99e7-2f71a602b110\") " pod="openstack/neutron-db-sync-2kr57" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.294104 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v8w7\" (UniqueName: \"kubernetes.io/projected/ab03b49b-be69-4e1d-99e7-2f71a602b110-kube-api-access-9v8w7\") pod \"neutron-db-sync-2kr57\" (UID: \"ab03b49b-be69-4e1d-99e7-2f71a602b110\") " pod="openstack/neutron-db-sync-2kr57" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.300338 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmp45\" (UniqueName: \"kubernetes.io/projected/d9c18809-9f62-4b7e-8103-df382e9b4fa6-kube-api-access-cmp45\") pod \"horizon-c455f8ff5-tkdz5\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.305832 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5ff9b486b9-79sqs"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.311621 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.339206 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-mgztm"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.340224 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mgztm" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.346814 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.347112 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.347215 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-2z6dv" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.359708 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-db-sync-config-data\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.359745 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-scripts\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.359789 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-config-data\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.359822 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-combined-ca-bundle\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.359841 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/04e3d721-2179-4457-8cc6-72af3b3f9257-etc-machine-id\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.359856 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9dbn\" (UniqueName: \"kubernetes.io/projected/04e3d721-2179-4457-8cc6-72af3b3f9257-kube-api-access-j9dbn\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.359890 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3643568-6595-4c73-9c83-be6e28fabe83-log-httpd\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.359908 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3643568-6595-4c73-9c83-be6e28fabe83-run-httpd\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.359933 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-scripts\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.359965 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mv9lz\" (UniqueName: \"kubernetes.io/projected/f3643568-6595-4c73-9c83-be6e28fabe83-kube-api-access-mv9lz\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.359989 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.360029 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.360056 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-config-data\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.363376 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3643568-6595-4c73-9c83-be6e28fabe83-run-httpd\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.363697 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3643568-6595-4c73-9c83-be6e28fabe83-log-httpd\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.374057 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.377890 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-scripts\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.379214 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-config-data\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.382987 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5ff9b486b9-79sqs"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.385173 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mv9lz\" (UniqueName: \"kubernetes.io/projected/f3643568-6595-4c73-9c83-be6e28fabe83-kube-api-access-mv9lz\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.385247 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.427427 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-mgztm"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.436493 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-lbk4j"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.438215 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.455625 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-lbk4j"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.461518 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-db-sync-config-data\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.461570 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-logs\") pod \"horizon-5ff9b486b9-79sqs\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.461619 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fzws\" (UniqueName: \"kubernetes.io/projected/ee89a5e7-e5a4-412f-a52a-511dfca61265-kube-api-access-7fzws\") pod \"placement-db-sync-mgztm\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " pod="openstack/placement-db-sync-mgztm" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.461645 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-combined-ca-bundle\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.461679 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/04e3d721-2179-4457-8cc6-72af3b3f9257-etc-machine-id\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.461700 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9dbn\" (UniqueName: \"kubernetes.io/projected/04e3d721-2179-4457-8cc6-72af3b3f9257-kube-api-access-j9dbn\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.461719 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee89a5e7-e5a4-412f-a52a-511dfca61265-logs\") pod \"placement-db-sync-mgztm\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " pod="openstack/placement-db-sync-mgztm" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.461743 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-config-data\") pod \"placement-db-sync-mgztm\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " pod="openstack/placement-db-sync-mgztm" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.461775 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-scripts\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.461790 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-scripts\") pod \"placement-db-sync-mgztm\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " pod="openstack/placement-db-sync-mgztm" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.461808 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-combined-ca-bundle\") pod \"placement-db-sync-mgztm\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " pod="openstack/placement-db-sync-mgztm" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.461850 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-horizon-secret-key\") pod \"horizon-5ff9b486b9-79sqs\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.461865 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2xhv\" (UniqueName: \"kubernetes.io/projected/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-kube-api-access-s2xhv\") pod \"horizon-5ff9b486b9-79sqs\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.461893 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-config-data\") pod \"horizon-5ff9b486b9-79sqs\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.461910 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-scripts\") pod \"horizon-5ff9b486b9-79sqs\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.461932 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-config-data\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.462878 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-d4ssg"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.462910 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/04e3d721-2179-4457-8cc6-72af3b3f9257-etc-machine-id\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.464021 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-d4ssg" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.466289 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mkf9f" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.466651 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.482542 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2kr57" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.484569 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-config-data\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.489008 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9dbn\" (UniqueName: \"kubernetes.io/projected/04e3d721-2179-4457-8cc6-72af3b3f9257-kube-api-access-j9dbn\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.502593 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-d4ssg"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.507671 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-db-sync-config-data\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.508383 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-combined-ca-bundle\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.515833 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.519243 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-scripts\") pod \"cinder-db-sync-576gs\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.568407 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.570633 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fzws\" (UniqueName: \"kubernetes.io/projected/ee89a5e7-e5a4-412f-a52a-511dfca61265-kube-api-access-7fzws\") pod \"placement-db-sync-mgztm\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " pod="openstack/placement-db-sync-mgztm" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.571324 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee89a5e7-e5a4-412f-a52a-511dfca61265-logs\") pod \"placement-db-sync-mgztm\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " pod="openstack/placement-db-sync-mgztm" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.571445 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-config-data\") pod \"placement-db-sync-mgztm\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " pod="openstack/placement-db-sync-mgztm" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.571572 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-scripts\") pod \"placement-db-sync-mgztm\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " pod="openstack/placement-db-sync-mgztm" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.572879 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.574732 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-combined-ca-bundle\") pod \"placement-db-sync-mgztm\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " pod="openstack/placement-db-sync-mgztm" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.574875 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.575002 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.575771 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-config\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.575941 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.572220 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee89a5e7-e5a4-412f-a52a-511dfca61265-logs\") pod \"placement-db-sync-mgztm\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " pod="openstack/placement-db-sync-mgztm" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.576261 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-horizon-secret-key\") pod \"horizon-5ff9b486b9-79sqs\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.576311 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2xhv\" (UniqueName: \"kubernetes.io/projected/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-kube-api-access-s2xhv\") pod \"horizon-5ff9b486b9-79sqs\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.576357 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz6rw\" (UniqueName: \"kubernetes.io/projected/73a1f5fc-5524-4753-9c6d-87c8d21f592a-kube-api-access-nz6rw\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.576440 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-config-data\") pod \"horizon-5ff9b486b9-79sqs\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.576469 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-scripts\") pod \"horizon-5ff9b486b9-79sqs\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.576602 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bdc9071-e863-4108-9234-4f201732f435-db-sync-config-data\") pod \"barbican-db-sync-d4ssg\" (UID: \"7bdc9071-e863-4108-9234-4f201732f435\") " pod="openstack/barbican-db-sync-d4ssg" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.576633 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-logs\") pod \"horizon-5ff9b486b9-79sqs\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.576684 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bdc9071-e863-4108-9234-4f201732f435-combined-ca-bundle\") pod \"barbican-db-sync-d4ssg\" (UID: \"7bdc9071-e863-4108-9234-4f201732f435\") " pod="openstack/barbican-db-sync-d4ssg" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.576722 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7spnb\" (UniqueName: \"kubernetes.io/projected/7bdc9071-e863-4108-9234-4f201732f435-kube-api-access-7spnb\") pod \"barbican-db-sync-d4ssg\" (UID: \"7bdc9071-e863-4108-9234-4f201732f435\") " pod="openstack/barbican-db-sync-d4ssg" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.578131 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-scripts\") pod \"placement-db-sync-mgztm\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " pod="openstack/placement-db-sync-mgztm" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.579311 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-config-data\") pod \"horizon-5ff9b486b9-79sqs\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.579911 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-scripts\") pod \"horizon-5ff9b486b9-79sqs\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.580236 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-logs\") pod \"horizon-5ff9b486b9-79sqs\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.584786 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-horizon-secret-key\") pod \"horizon-5ff9b486b9-79sqs\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.591141 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-combined-ca-bundle\") pod \"placement-db-sync-mgztm\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " pod="openstack/placement-db-sync-mgztm" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.592234 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-576gs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.601428 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.602913 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-config-data\") pod \"placement-db-sync-mgztm\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " pod="openstack/placement-db-sync-mgztm" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.607211 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.626064 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-dqnk7" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.626417 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.626789 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.627038 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.627105 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fzws\" (UniqueName: \"kubernetes.io/projected/ee89a5e7-e5a4-412f-a52a-511dfca61265-kube-api-access-7fzws\") pod \"placement-db-sync-mgztm\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " pod="openstack/placement-db-sync-mgztm" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.627493 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2xhv\" (UniqueName: \"kubernetes.io/projected/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-kube-api-access-s2xhv\") pod \"horizon-5ff9b486b9-79sqs\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.632983 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.639100 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.677619 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz6rw\" (UniqueName: \"kubernetes.io/projected/73a1f5fc-5524-4753-9c6d-87c8d21f592a-kube-api-access-nz6rw\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.677779 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bdc9071-e863-4108-9234-4f201732f435-db-sync-config-data\") pod \"barbican-db-sync-d4ssg\" (UID: \"7bdc9071-e863-4108-9234-4f201732f435\") " pod="openstack/barbican-db-sync-d4ssg" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.677859 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bdc9071-e863-4108-9234-4f201732f435-combined-ca-bundle\") pod \"barbican-db-sync-d4ssg\" (UID: \"7bdc9071-e863-4108-9234-4f201732f435\") " pod="openstack/barbican-db-sync-d4ssg" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.677926 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7spnb\" (UniqueName: \"kubernetes.io/projected/7bdc9071-e863-4108-9234-4f201732f435-kube-api-access-7spnb\") pod \"barbican-db-sync-d4ssg\" (UID: \"7bdc9071-e863-4108-9234-4f201732f435\") " pod="openstack/barbican-db-sync-d4ssg" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.678033 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.678102 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.678274 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.678344 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-config\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.678408 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.679229 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.680180 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.680941 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.681491 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.685304 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mgztm" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.690864 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bdc9071-e863-4108-9234-4f201732f435-combined-ca-bundle\") pod \"barbican-db-sync-d4ssg\" (UID: \"7bdc9071-e863-4108-9234-4f201732f435\") " pod="openstack/barbican-db-sync-d4ssg" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.691047 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bdc9071-e863-4108-9234-4f201732f435-db-sync-config-data\") pod \"barbican-db-sync-d4ssg\" (UID: \"7bdc9071-e863-4108-9234-4f201732f435\") " pod="openstack/barbican-db-sync-d4ssg" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.691615 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-config\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.701489 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz6rw\" (UniqueName: \"kubernetes.io/projected/73a1f5fc-5524-4753-9c6d-87c8d21f592a-kube-api-access-nz6rw\") pod \"dnsmasq-dns-57c957c4ff-lbk4j\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.707183 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7spnb\" (UniqueName: \"kubernetes.io/projected/7bdc9071-e863-4108-9234-4f201732f435-kube-api-access-7spnb\") pod \"barbican-db-sync-d4ssg\" (UID: \"7bdc9071-e863-4108-9234-4f201732f435\") " pod="openstack/barbican-db-sync-d4ssg" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.780247 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-config-data\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.780292 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/baf9a094-aa08-4303-bd7f-4e661063f82e-logs\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.780374 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.780425 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/baf9a094-aa08-4303-bd7f-4e661063f82e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.780451 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-scripts\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.780492 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.780521 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvz8c\" (UniqueName: \"kubernetes.io/projected/baf9a094-aa08-4303-bd7f-4e661063f82e-kube-api-access-dvz8c\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.780608 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.880831 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-chzbc"] Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.882062 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.882089 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/baf9a094-aa08-4303-bd7f-4e661063f82e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.882108 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-scripts\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.882129 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.882145 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvz8c\" (UniqueName: \"kubernetes.io/projected/baf9a094-aa08-4303-bd7f-4e661063f82e-kube-api-access-dvz8c\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.882206 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.882244 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-config-data\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.882264 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/baf9a094-aa08-4303-bd7f-4e661063f82e-logs\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.882848 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/baf9a094-aa08-4303-bd7f-4e661063f82e-logs\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.883164 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.884163 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/baf9a094-aa08-4303-bd7f-4e661063f82e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.884215 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.888431 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-config-data\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.889132 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.889937 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-scripts\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.892795 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.903313 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvz8c\" (UniqueName: \"kubernetes.io/projected/baf9a094-aa08-4303-bd7f-4e661063f82e-kube-api-access-dvz8c\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.915550 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:44 crc kubenswrapper[4815]: W0225 13:37:44.916659 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8df64e8_8db3_4a60_807d_f8cf502320e4.slice/crio-e0750e4893d0a4055ea57ce27aba0c391d99b571c19c2b0297e89b2630647513 WatchSource:0}: Error finding container e0750e4893d0a4055ea57ce27aba0c391d99b571c19c2b0297e89b2630647513: Status 404 returned error can't find the container with id e0750e4893d0a4055ea57ce27aba0c391d99b571c19c2b0297e89b2630647513 Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.927983 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-d4ssg" Feb 25 13:37:44 crc kubenswrapper[4815]: I0225 13:37:44.979588 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.099397 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2kr57"] Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.106371 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.108141 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.111744 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.112959 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.157248 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.228052 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-qs8h9"] Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.237487 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c455f8ff5-tkdz5"] Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.293584 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.293785 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26e0f670-91b0-4bed-8a45-596bcf916ce3-logs\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.293864 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.294205 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.294331 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26e0f670-91b0-4bed-8a45-596bcf916ce3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.294433 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.294564 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.295607 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t7qn\" (UniqueName: \"kubernetes.io/projected/26e0f670-91b0-4bed-8a45-596bcf916ce3-kube-api-access-2t7qn\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.375499 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.403614 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.403654 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26e0f670-91b0-4bed-8a45-596bcf916ce3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.403680 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.403704 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.403721 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t7qn\" (UniqueName: \"kubernetes.io/projected/26e0f670-91b0-4bed-8a45-596bcf916ce3-kube-api-access-2t7qn\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.403779 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.403803 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26e0f670-91b0-4bed-8a45-596bcf916ce3-logs\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.403818 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.403964 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.405537 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26e0f670-91b0-4bed-8a45-596bcf916ce3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.405805 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26e0f670-91b0-4bed-8a45-596bcf916ce3-logs\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.422222 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.433683 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.436216 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.437453 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.452134 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t7qn\" (UniqueName: \"kubernetes.io/projected/26e0f670-91b0-4bed-8a45-596bcf916ce3-kube-api-access-2t7qn\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.478546 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c455f8ff5-tkdz5" event={"ID":"d9c18809-9f62-4b7e-8103-df382e9b4fa6","Type":"ContainerStarted","Data":"50acff9ff504baddfe9194e3d3802a75cd3bfd040c6efa1bf77167cea43ec545"} Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.482148 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.494189 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3643568-6595-4c73-9c83-be6e28fabe83","Type":"ContainerStarted","Data":"7f8fd4524137989ea9b83636c13dfa58a0ca76969ae71f4c03aab4111d99265f"} Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.500012 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qs8h9" event={"ID":"c36a30b8-6a56-4f23-8c5e-7556199b2ad4","Type":"ContainerStarted","Data":"d2218c3e8b1f23653d1a708ab1f4275aa530220b625011e8908d14cf8b45bec8"} Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.501245 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" event={"ID":"e8df64e8-8db3-4a60-807d-f8cf502320e4","Type":"ContainerStarted","Data":"e0750e4893d0a4055ea57ce27aba0c391d99b571c19c2b0297e89b2630647513"} Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.502445 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2kr57" event={"ID":"ab03b49b-be69-4e1d-99e7-2f71a602b110","Type":"ContainerStarted","Data":"b2208242ae9be094b621e23d25975615ccd3e0c5acf428274eb4238accb00bc8"} Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.584872 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.612018 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-mgztm"] Feb 25 13:37:45 crc kubenswrapper[4815]: W0225 13:37:45.657682 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee89a5e7_e5a4_412f_a52a_511dfca61265.slice/crio-c57f5520a69df9e29d9777d5cfc7db02cf80f6e84cdc3d18c36b829306d04eb8 WatchSource:0}: Error finding container c57f5520a69df9e29d9777d5cfc7db02cf80f6e84cdc3d18c36b829306d04eb8: Status 404 returned error can't find the container with id c57f5520a69df9e29d9777d5cfc7db02cf80f6e84cdc3d18c36b829306d04eb8 Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.713807 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5ff9b486b9-79sqs"] Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.724736 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-576gs"] Feb 25 13:37:45 crc kubenswrapper[4815]: I0225 13:37:45.777356 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-lbk4j"] Feb 25 13:37:45 crc kubenswrapper[4815]: W0225 13:37:45.784706 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73a1f5fc_5524_4753_9c6d_87c8d21f592a.slice/crio-6f3785a779335d37182644a434564d3b739abdacced322230a12f2aec88fd60c WatchSource:0}: Error finding container 6f3785a779335d37182644a434564d3b739abdacced322230a12f2aec88fd60c: Status 404 returned error can't find the container with id 6f3785a779335d37182644a434564d3b739abdacced322230a12f2aec88fd60c Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.011765 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-d4ssg"] Feb 25 13:37:46 crc kubenswrapper[4815]: W0225 13:37:46.060582 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bdc9071_e863_4108_9234_4f201732f435.slice/crio-fbbd68a725326ce247e9bca7810965261c130ba37fa09d8cc74505e052af8d96 WatchSource:0}: Error finding container fbbd68a725326ce247e9bca7810965261c130ba37fa09d8cc74505e052af8d96: Status 404 returned error can't find the container with id fbbd68a725326ce247e9bca7810965261c130ba37fa09d8cc74505e052af8d96 Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.112250 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:37:46 crc kubenswrapper[4815]: W0225 13:37:46.183104 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbaf9a094_aa08_4303_bd7f_4e661063f82e.slice/crio-e02614ccda98558acf80dfba6734400008c66b4a242a03f89f3f5a176d0fe9dc WatchSource:0}: Error finding container e02614ccda98558acf80dfba6734400008c66b4a242a03f89f3f5a176d0fe9dc: Status 404 returned error can't find the container with id e02614ccda98558acf80dfba6734400008c66b4a242a03f89f3f5a176d0fe9dc Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.239065 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.345176 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.389176 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c455f8ff5-tkdz5"] Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.439129 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-85878dd5b5-q9gfx"] Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.440683 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.452126 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.472921 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.492563 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85878dd5b5-q9gfx"] Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.524751 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-576gs" event={"ID":"04e3d721-2179-4457-8cc6-72af3b3f9257","Type":"ContainerStarted","Data":"24b242a8edf5fb48d59efcd722a53de16f270444b84375f434536b74320b3ed7"} Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.536995 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-logs\") pod \"horizon-85878dd5b5-q9gfx\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.537909 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-scripts\") pod \"horizon-85878dd5b5-q9gfx\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.538069 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-horizon-secret-key\") pod \"horizon-85878dd5b5-q9gfx\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.538163 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhhln\" (UniqueName: \"kubernetes.io/projected/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-kube-api-access-mhhln\") pod \"horizon-85878dd5b5-q9gfx\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.538232 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-config-data\") pod \"horizon-85878dd5b5-q9gfx\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.544585 4815 generic.go:334] "Generic (PLEG): container finished" podID="73a1f5fc-5524-4753-9c6d-87c8d21f592a" containerID="72af52454ed4764867fb2a872f2680abe7428323592a8731116c568dbf259e4d" exitCode=0 Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.545628 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" event={"ID":"73a1f5fc-5524-4753-9c6d-87c8d21f592a","Type":"ContainerDied","Data":"72af52454ed4764867fb2a872f2680abe7428323592a8731116c568dbf259e4d"} Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.545660 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" event={"ID":"73a1f5fc-5524-4753-9c6d-87c8d21f592a","Type":"ContainerStarted","Data":"6f3785a779335d37182644a434564d3b739abdacced322230a12f2aec88fd60c"} Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.571581 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mgztm" event={"ID":"ee89a5e7-e5a4-412f-a52a-511dfca61265","Type":"ContainerStarted","Data":"c57f5520a69df9e29d9777d5cfc7db02cf80f6e84cdc3d18c36b829306d04eb8"} Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.590702 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5ff9b486b9-79sqs" event={"ID":"ef5ee7f2-aefa-4107-8fa9-57902643ab8e","Type":"ContainerStarted","Data":"6707fce2a9afc6ed1a5c272214047fa2a2849dbc68b46d6f84b3f659ee0652ce"} Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.592353 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-d4ssg" event={"ID":"7bdc9071-e863-4108-9234-4f201732f435","Type":"ContainerStarted","Data":"fbbd68a725326ce247e9bca7810965261c130ba37fa09d8cc74505e052af8d96"} Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.594569 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"baf9a094-aa08-4303-bd7f-4e661063f82e","Type":"ContainerStarted","Data":"e02614ccda98558acf80dfba6734400008c66b4a242a03f89f3f5a176d0fe9dc"} Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.600843 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2kr57" event={"ID":"ab03b49b-be69-4e1d-99e7-2f71a602b110","Type":"ContainerStarted","Data":"4dd4c4d4dba488e7472dcc7279f6c3ad41d76de7b59eb5521533e41a8e0c8a40"} Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.620851 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"26e0f670-91b0-4bed-8a45-596bcf916ce3","Type":"ContainerStarted","Data":"28f8e62b0377f288e99cdd111648bbc41d158ac25445d742700df51f6cdb097a"} Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.624571 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-2kr57" podStartSLOduration=3.624538404 podStartE2EDuration="3.624538404s" podCreationTimestamp="2026-02-25 13:37:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:37:46.617206768 +0000 UTC m=+1024.418304812" watchObservedRunningTime="2026-02-25 13:37:46.624538404 +0000 UTC m=+1024.425636458" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.627965 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qs8h9" event={"ID":"c36a30b8-6a56-4f23-8c5e-7556199b2ad4","Type":"ContainerStarted","Data":"5d82924ff0f902481efdf03a6e242bf482ae47b4379d98e8299ecdfbb0c47469"} Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.630725 4815 generic.go:334] "Generic (PLEG): container finished" podID="e8df64e8-8db3-4a60-807d-f8cf502320e4" containerID="bd5ea23912bd0d89088d344bbe49f47673a47b24f5ae66835b996578b3816b79" exitCode=0 Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.630785 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" event={"ID":"e8df64e8-8db3-4a60-807d-f8cf502320e4","Type":"ContainerDied","Data":"bd5ea23912bd0d89088d344bbe49f47673a47b24f5ae66835b996578b3816b79"} Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.639920 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-logs\") pod \"horizon-85878dd5b5-q9gfx\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.640001 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-scripts\") pod \"horizon-85878dd5b5-q9gfx\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.640045 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-horizon-secret-key\") pod \"horizon-85878dd5b5-q9gfx\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.640070 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhhln\" (UniqueName: \"kubernetes.io/projected/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-kube-api-access-mhhln\") pod \"horizon-85878dd5b5-q9gfx\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.640096 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-config-data\") pod \"horizon-85878dd5b5-q9gfx\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.646670 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-qs8h9" podStartSLOduration=3.646655557 podStartE2EDuration="3.646655557s" podCreationTimestamp="2026-02-25 13:37:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:37:46.644845032 +0000 UTC m=+1024.445943086" watchObservedRunningTime="2026-02-25 13:37:46.646655557 +0000 UTC m=+1024.447753611" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.648209 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-scripts\") pod \"horizon-85878dd5b5-q9gfx\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.646690 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-logs\") pod \"horizon-85878dd5b5-q9gfx\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.648739 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-config-data\") pod \"horizon-85878dd5b5-q9gfx\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.663411 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-horizon-secret-key\") pod \"horizon-85878dd5b5-q9gfx\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.667645 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhhln\" (UniqueName: \"kubernetes.io/projected/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-kube-api-access-mhhln\") pod \"horizon-85878dd5b5-q9gfx\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:37:46 crc kubenswrapper[4815]: I0225 13:37:46.781915 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.231613 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.375864 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xkm4\" (UniqueName: \"kubernetes.io/projected/e8df64e8-8db3-4a60-807d-f8cf502320e4-kube-api-access-2xkm4\") pod \"e8df64e8-8db3-4a60-807d-f8cf502320e4\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.375972 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-config\") pod \"e8df64e8-8db3-4a60-807d-f8cf502320e4\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.375996 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-dns-svc\") pod \"e8df64e8-8db3-4a60-807d-f8cf502320e4\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.376026 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-dns-swift-storage-0\") pod \"e8df64e8-8db3-4a60-807d-f8cf502320e4\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.376361 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-ovsdbserver-nb\") pod \"e8df64e8-8db3-4a60-807d-f8cf502320e4\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.376733 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-ovsdbserver-sb\") pod \"e8df64e8-8db3-4a60-807d-f8cf502320e4\" (UID: \"e8df64e8-8db3-4a60-807d-f8cf502320e4\") " Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.379300 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8df64e8-8db3-4a60-807d-f8cf502320e4-kube-api-access-2xkm4" (OuterVolumeSpecName: "kube-api-access-2xkm4") pod "e8df64e8-8db3-4a60-807d-f8cf502320e4" (UID: "e8df64e8-8db3-4a60-807d-f8cf502320e4"). InnerVolumeSpecName "kube-api-access-2xkm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.411496 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-config" (OuterVolumeSpecName: "config") pod "e8df64e8-8db3-4a60-807d-f8cf502320e4" (UID: "e8df64e8-8db3-4a60-807d-f8cf502320e4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.411524 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e8df64e8-8db3-4a60-807d-f8cf502320e4" (UID: "e8df64e8-8db3-4a60-807d-f8cf502320e4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.419202 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e8df64e8-8db3-4a60-807d-f8cf502320e4" (UID: "e8df64e8-8db3-4a60-807d-f8cf502320e4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.423268 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e8df64e8-8db3-4a60-807d-f8cf502320e4" (UID: "e8df64e8-8db3-4a60-807d-f8cf502320e4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.441930 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85878dd5b5-q9gfx"] Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.451002 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e8df64e8-8db3-4a60-807d-f8cf502320e4" (UID: "e8df64e8-8db3-4a60-807d-f8cf502320e4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:37:47 crc kubenswrapper[4815]: W0225 13:37:47.457600 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod755e6fea_b2e0_4dfb_99e9_6ecad78e754d.slice/crio-f09c98a586f8643d63288327c5a64153cd634fda881aac812423d2b992ea46c9 WatchSource:0}: Error finding container f09c98a586f8643d63288327c5a64153cd634fda881aac812423d2b992ea46c9: Status 404 returned error can't find the container with id f09c98a586f8643d63288327c5a64153cd634fda881aac812423d2b992ea46c9 Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.484516 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.484823 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xkm4\" (UniqueName: \"kubernetes.io/projected/e8df64e8-8db3-4a60-807d-f8cf502320e4-kube-api-access-2xkm4\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.484833 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.484842 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.484852 4815 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.484860 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8df64e8-8db3-4a60-807d-f8cf502320e4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.733243 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"baf9a094-aa08-4303-bd7f-4e661063f82e","Type":"ContainerStarted","Data":"fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89"} Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.757955 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" event={"ID":"73a1f5fc-5524-4753-9c6d-87c8d21f592a","Type":"ContainerStarted","Data":"5ad820c001b1adaa66f858b1dff7082c42d1a32530cbe5680e51e85725e39f48"} Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.758909 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.769673 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85878dd5b5-q9gfx" event={"ID":"755e6fea-b2e0-4dfb-99e9-6ecad78e754d","Type":"ContainerStarted","Data":"f09c98a586f8643d63288327c5a64153cd634fda881aac812423d2b992ea46c9"} Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.774471 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"26e0f670-91b0-4bed-8a45-596bcf916ce3","Type":"ContainerStarted","Data":"c6a74f1c0e7af9d7ae823c955b720f65bcea3698261db790707090128b3f9024"} Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.777221 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.783695 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c9f998c-chzbc" event={"ID":"e8df64e8-8db3-4a60-807d-f8cf502320e4","Type":"ContainerDied","Data":"e0750e4893d0a4055ea57ce27aba0c391d99b571c19c2b0297e89b2630647513"} Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.783737 4815 scope.go:117] "RemoveContainer" containerID="bd5ea23912bd0d89088d344bbe49f47673a47b24f5ae66835b996578b3816b79" Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.796195 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" podStartSLOduration=3.796179776 podStartE2EDuration="3.796179776s" podCreationTimestamp="2026-02-25 13:37:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:37:47.791904254 +0000 UTC m=+1025.593002308" watchObservedRunningTime="2026-02-25 13:37:47.796179776 +0000 UTC m=+1025.597277830" Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.874456 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-chzbc"] Feb 25 13:37:47 crc kubenswrapper[4815]: I0225 13:37:47.885057 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-chzbc"] Feb 25 13:37:48 crc kubenswrapper[4815]: I0225 13:37:48.808334 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"baf9a094-aa08-4303-bd7f-4e661063f82e","Type":"ContainerStarted","Data":"d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5"} Feb 25 13:37:48 crc kubenswrapper[4815]: I0225 13:37:48.808764 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="baf9a094-aa08-4303-bd7f-4e661063f82e" containerName="glance-log" containerID="cri-o://fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89" gracePeriod=30 Feb 25 13:37:48 crc kubenswrapper[4815]: I0225 13:37:48.809003 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="baf9a094-aa08-4303-bd7f-4e661063f82e" containerName="glance-httpd" containerID="cri-o://d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5" gracePeriod=30 Feb 25 13:37:48 crc kubenswrapper[4815]: I0225 13:37:48.820699 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="26e0f670-91b0-4bed-8a45-596bcf916ce3" containerName="glance-log" containerID="cri-o://c6a74f1c0e7af9d7ae823c955b720f65bcea3698261db790707090128b3f9024" gracePeriod=30 Feb 25 13:37:48 crc kubenswrapper[4815]: I0225 13:37:48.820870 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="26e0f670-91b0-4bed-8a45-596bcf916ce3" containerName="glance-httpd" containerID="cri-o://7df816bfd022b3c0e86bc9e077e3a98903d2b2854df2d805581f0efa43ba685f" gracePeriod=30 Feb 25 13:37:48 crc kubenswrapper[4815]: I0225 13:37:48.821092 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"26e0f670-91b0-4bed-8a45-596bcf916ce3","Type":"ContainerStarted","Data":"7df816bfd022b3c0e86bc9e077e3a98903d2b2854df2d805581f0efa43ba685f"} Feb 25 13:37:48 crc kubenswrapper[4815]: I0225 13:37:48.836944 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.836921923 podStartE2EDuration="4.836921923s" podCreationTimestamp="2026-02-25 13:37:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:37:48.833417494 +0000 UTC m=+1026.634515548" watchObservedRunningTime="2026-02-25 13:37:48.836921923 +0000 UTC m=+1026.638019977" Feb 25 13:37:48 crc kubenswrapper[4815]: I0225 13:37:48.863379 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.863361589 podStartE2EDuration="4.863361589s" podCreationTimestamp="2026-02-25 13:37:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:37:48.858709236 +0000 UTC m=+1026.659807290" watchObservedRunningTime="2026-02-25 13:37:48.863361589 +0000 UTC m=+1026.664459643" Feb 25 13:37:48 crc kubenswrapper[4815]: I0225 13:37:48.956297 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8df64e8-8db3-4a60-807d-f8cf502320e4" path="/var/lib/kubelet/pods/e8df64e8-8db3-4a60-807d-f8cf502320e4/volumes" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.569601 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.737842 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-combined-ca-bundle\") pod \"baf9a094-aa08-4303-bd7f-4e661063f82e\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.737949 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/baf9a094-aa08-4303-bd7f-4e661063f82e-logs\") pod \"baf9a094-aa08-4303-bd7f-4e661063f82e\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.738002 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-scripts\") pod \"baf9a094-aa08-4303-bd7f-4e661063f82e\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.738035 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-public-tls-certs\") pod \"baf9a094-aa08-4303-bd7f-4e661063f82e\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.738063 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvz8c\" (UniqueName: \"kubernetes.io/projected/baf9a094-aa08-4303-bd7f-4e661063f82e-kube-api-access-dvz8c\") pod \"baf9a094-aa08-4303-bd7f-4e661063f82e\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.738079 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/baf9a094-aa08-4303-bd7f-4e661063f82e-httpd-run\") pod \"baf9a094-aa08-4303-bd7f-4e661063f82e\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.738148 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"baf9a094-aa08-4303-bd7f-4e661063f82e\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.738189 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-config-data\") pod \"baf9a094-aa08-4303-bd7f-4e661063f82e\" (UID: \"baf9a094-aa08-4303-bd7f-4e661063f82e\") " Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.738907 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/baf9a094-aa08-4303-bd7f-4e661063f82e-logs" (OuterVolumeSpecName: "logs") pod "baf9a094-aa08-4303-bd7f-4e661063f82e" (UID: "baf9a094-aa08-4303-bd7f-4e661063f82e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.738925 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/baf9a094-aa08-4303-bd7f-4e661063f82e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "baf9a094-aa08-4303-bd7f-4e661063f82e" (UID: "baf9a094-aa08-4303-bd7f-4e661063f82e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.747986 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-scripts" (OuterVolumeSpecName: "scripts") pod "baf9a094-aa08-4303-bd7f-4e661063f82e" (UID: "baf9a094-aa08-4303-bd7f-4e661063f82e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.756035 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "baf9a094-aa08-4303-bd7f-4e661063f82e" (UID: "baf9a094-aa08-4303-bd7f-4e661063f82e"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.757733 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baf9a094-aa08-4303-bd7f-4e661063f82e-kube-api-access-dvz8c" (OuterVolumeSpecName: "kube-api-access-dvz8c") pod "baf9a094-aa08-4303-bd7f-4e661063f82e" (UID: "baf9a094-aa08-4303-bd7f-4e661063f82e"). InnerVolumeSpecName "kube-api-access-dvz8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.803191 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-config-data" (OuterVolumeSpecName: "config-data") pod "baf9a094-aa08-4303-bd7f-4e661063f82e" (UID: "baf9a094-aa08-4303-bd7f-4e661063f82e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.815558 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "baf9a094-aa08-4303-bd7f-4e661063f82e" (UID: "baf9a094-aa08-4303-bd7f-4e661063f82e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.815657 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "baf9a094-aa08-4303-bd7f-4e661063f82e" (UID: "baf9a094-aa08-4303-bd7f-4e661063f82e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.831279 4815 generic.go:334] "Generic (PLEG): container finished" podID="baf9a094-aa08-4303-bd7f-4e661063f82e" containerID="d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5" exitCode=0 Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.831316 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"baf9a094-aa08-4303-bd7f-4e661063f82e","Type":"ContainerDied","Data":"d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5"} Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.831353 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"baf9a094-aa08-4303-bd7f-4e661063f82e","Type":"ContainerDied","Data":"fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89"} Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.831295 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.831313 4815 generic.go:334] "Generic (PLEG): container finished" podID="baf9a094-aa08-4303-bd7f-4e661063f82e" containerID="fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89" exitCode=143 Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.831444 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"baf9a094-aa08-4303-bd7f-4e661063f82e","Type":"ContainerDied","Data":"e02614ccda98558acf80dfba6734400008c66b4a242a03f89f3f5a176d0fe9dc"} Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.831484 4815 scope.go:117] "RemoveContainer" containerID="d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.840558 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.840584 4815 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.840595 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvz8c\" (UniqueName: \"kubernetes.io/projected/baf9a094-aa08-4303-bd7f-4e661063f82e-kube-api-access-dvz8c\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.840605 4815 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/baf9a094-aa08-4303-bd7f-4e661063f82e-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.840633 4815 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.840644 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.840652 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf9a094-aa08-4303-bd7f-4e661063f82e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.840661 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/baf9a094-aa08-4303-bd7f-4e661063f82e-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.856140 4815 generic.go:334] "Generic (PLEG): container finished" podID="26e0f670-91b0-4bed-8a45-596bcf916ce3" containerID="7df816bfd022b3c0e86bc9e077e3a98903d2b2854df2d805581f0efa43ba685f" exitCode=0 Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.856171 4815 generic.go:334] "Generic (PLEG): container finished" podID="26e0f670-91b0-4bed-8a45-596bcf916ce3" containerID="c6a74f1c0e7af9d7ae823c955b720f65bcea3698261db790707090128b3f9024" exitCode=143 Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.856192 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"26e0f670-91b0-4bed-8a45-596bcf916ce3","Type":"ContainerDied","Data":"7df816bfd022b3c0e86bc9e077e3a98903d2b2854df2d805581f0efa43ba685f"} Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.856228 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"26e0f670-91b0-4bed-8a45-596bcf916ce3","Type":"ContainerDied","Data":"c6a74f1c0e7af9d7ae823c955b720f65bcea3698261db790707090128b3f9024"} Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.859656 4815 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.873488 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.881552 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.893202 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:37:49 crc kubenswrapper[4815]: E0225 13:37:49.895144 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baf9a094-aa08-4303-bd7f-4e661063f82e" containerName="glance-log" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.895163 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="baf9a094-aa08-4303-bd7f-4e661063f82e" containerName="glance-log" Feb 25 13:37:49 crc kubenswrapper[4815]: E0225 13:37:49.895193 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8df64e8-8db3-4a60-807d-f8cf502320e4" containerName="init" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.895199 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8df64e8-8db3-4a60-807d-f8cf502320e4" containerName="init" Feb 25 13:37:49 crc kubenswrapper[4815]: E0225 13:37:49.895216 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baf9a094-aa08-4303-bd7f-4e661063f82e" containerName="glance-httpd" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.895222 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="baf9a094-aa08-4303-bd7f-4e661063f82e" containerName="glance-httpd" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.895400 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8df64e8-8db3-4a60-807d-f8cf502320e4" containerName="init" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.895419 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="baf9a094-aa08-4303-bd7f-4e661063f82e" containerName="glance-httpd" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.895433 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="baf9a094-aa08-4303-bd7f-4e661063f82e" containerName="glance-log" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.896528 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.906004 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.907188 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.912894 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:37:49 crc kubenswrapper[4815]: I0225 13:37:49.942732 4815 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.044251 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-scripts\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.044287 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.044318 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49be5892-2d7d-4e75-bf93-6a1d16e769fe-logs\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.044336 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df9xt\" (UniqueName: \"kubernetes.io/projected/49be5892-2d7d-4e75-bf93-6a1d16e769fe-kube-api-access-df9xt\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.044381 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.044413 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.044458 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-config-data\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.044527 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/49be5892-2d7d-4e75-bf93-6a1d16e769fe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.146077 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-scripts\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.146118 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.146154 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49be5892-2d7d-4e75-bf93-6a1d16e769fe-logs\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.146174 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df9xt\" (UniqueName: \"kubernetes.io/projected/49be5892-2d7d-4e75-bf93-6a1d16e769fe-kube-api-access-df9xt\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.146214 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.146247 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.146292 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-config-data\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.146339 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/49be5892-2d7d-4e75-bf93-6a1d16e769fe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.147404 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/49be5892-2d7d-4e75-bf93-6a1d16e769fe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.147773 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.149089 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49be5892-2d7d-4e75-bf93-6a1d16e769fe-logs\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.150957 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.160020 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.160658 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-scripts\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.165208 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-config-data\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.166145 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df9xt\" (UniqueName: \"kubernetes.io/projected/49be5892-2d7d-4e75-bf93-6a1d16e769fe-kube-api-access-df9xt\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.235557 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.257225 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.865687 4815 generic.go:334] "Generic (PLEG): container finished" podID="c36a30b8-6a56-4f23-8c5e-7556199b2ad4" containerID="5d82924ff0f902481efdf03a6e242bf482ae47b4379d98e8299ecdfbb0c47469" exitCode=0 Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.865755 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qs8h9" event={"ID":"c36a30b8-6a56-4f23-8c5e-7556199b2ad4","Type":"ContainerDied","Data":"5d82924ff0f902481efdf03a6e242bf482ae47b4379d98e8299ecdfbb0c47469"} Feb 25 13:37:50 crc kubenswrapper[4815]: I0225 13:37:50.946959 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baf9a094-aa08-4303-bd7f-4e661063f82e" path="/var/lib/kubelet/pods/baf9a094-aa08-4303-bd7f-4e661063f82e/volumes" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.142447 4815 scope.go:117] "RemoveContainer" containerID="fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.468926 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5ff9b486b9-79sqs"] Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.496889 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6497d456d-s5g59"] Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.498212 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.500427 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.543420 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6497d456d-s5g59"] Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.591700 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.608011 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-horizon-tls-certs\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.608116 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-horizon-secret-key\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.608140 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4jm8\" (UniqueName: \"kubernetes.io/projected/5d076fbb-4499-456f-82ee-211a8909fc99-kube-api-access-r4jm8\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.608172 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d076fbb-4499-456f-82ee-211a8909fc99-logs\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.608202 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d076fbb-4499-456f-82ee-211a8909fc99-config-data\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.608230 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-combined-ca-bundle\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.608246 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d076fbb-4499-456f-82ee-211a8909fc99-scripts\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.641810 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85878dd5b5-q9gfx"] Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.652801 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7657655688-k9k6t"] Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.654304 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.665447 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7657655688-k9k6t"] Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.709436 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4jm8\" (UniqueName: \"kubernetes.io/projected/5d076fbb-4499-456f-82ee-211a8909fc99-kube-api-access-r4jm8\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.709483 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f10b8779-9d5e-427d-847c-3250816e89b9-horizon-secret-key\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.709548 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f10b8779-9d5e-427d-847c-3250816e89b9-config-data\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.709574 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d076fbb-4499-456f-82ee-211a8909fc99-logs\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.709721 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10b8779-9d5e-427d-847c-3250816e89b9-combined-ca-bundle\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.709794 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d076fbb-4499-456f-82ee-211a8909fc99-config-data\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.709876 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-combined-ca-bundle\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.709921 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f10b8779-9d5e-427d-847c-3250816e89b9-scripts\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.709929 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d076fbb-4499-456f-82ee-211a8909fc99-logs\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.709949 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d076fbb-4499-456f-82ee-211a8909fc99-scripts\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.710041 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f10b8779-9d5e-427d-847c-3250816e89b9-logs\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.710104 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-horizon-tls-certs\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.710162 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhc49\" (UniqueName: \"kubernetes.io/projected/f10b8779-9d5e-427d-847c-3250816e89b9-kube-api-access-mhc49\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.710558 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f10b8779-9d5e-427d-847c-3250816e89b9-horizon-tls-certs\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.710585 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-horizon-secret-key\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.711439 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d076fbb-4499-456f-82ee-211a8909fc99-config-data\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.711641 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d076fbb-4499-456f-82ee-211a8909fc99-scripts\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.720028 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-horizon-tls-certs\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.720027 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-combined-ca-bundle\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.730346 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4jm8\" (UniqueName: \"kubernetes.io/projected/5d076fbb-4499-456f-82ee-211a8909fc99-kube-api-access-r4jm8\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.742948 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-horizon-secret-key\") pod \"horizon-6497d456d-s5g59\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.812124 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f10b8779-9d5e-427d-847c-3250816e89b9-config-data\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.812185 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10b8779-9d5e-427d-847c-3250816e89b9-combined-ca-bundle\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.812242 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f10b8779-9d5e-427d-847c-3250816e89b9-scripts\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.812290 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f10b8779-9d5e-427d-847c-3250816e89b9-logs\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.812334 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhc49\" (UniqueName: \"kubernetes.io/projected/f10b8779-9d5e-427d-847c-3250816e89b9-kube-api-access-mhc49\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.812417 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f10b8779-9d5e-427d-847c-3250816e89b9-horizon-tls-certs\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.812455 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f10b8779-9d5e-427d-847c-3250816e89b9-horizon-secret-key\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.813218 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f10b8779-9d5e-427d-847c-3250816e89b9-logs\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.813274 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f10b8779-9d5e-427d-847c-3250816e89b9-scripts\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.813343 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f10b8779-9d5e-427d-847c-3250816e89b9-config-data\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.817087 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f10b8779-9d5e-427d-847c-3250816e89b9-horizon-tls-certs\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.817575 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f10b8779-9d5e-427d-847c-3250816e89b9-horizon-secret-key\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.827004 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10b8779-9d5e-427d-847c-3250816e89b9-combined-ca-bundle\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.835106 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhc49\" (UniqueName: \"kubernetes.io/projected/f10b8779-9d5e-427d-847c-3250816e89b9-kube-api-access-mhc49\") pod \"horizon-7657655688-k9k6t\" (UID: \"f10b8779-9d5e-427d-847c-3250816e89b9\") " pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.836823 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:37:53 crc kubenswrapper[4815]: I0225 13:37:53.978709 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:37:54 crc kubenswrapper[4815]: I0225 13:37:54.885794 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:37:54 crc kubenswrapper[4815]: I0225 13:37:54.978293 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-5lnz6"] Feb 25 13:37:54 crc kubenswrapper[4815]: I0225 13:37:54.978541 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" podUID="3199ac65-2fd7-482f-b2fa-b5305ffa59d8" containerName="dnsmasq-dns" containerID="cri-o://35b893c825ea2cf30ddd270904b104304eed3db2ef55461fe4ab5b3edeb5c804" gracePeriod=10 Feb 25 13:37:55 crc kubenswrapper[4815]: I0225 13:37:55.807434 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" podUID="3199ac65-2fd7-482f-b2fa-b5305ffa59d8" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: connect: connection refused" Feb 25 13:37:55 crc kubenswrapper[4815]: I0225 13:37:55.927965 4815 generic.go:334] "Generic (PLEG): container finished" podID="3199ac65-2fd7-482f-b2fa-b5305ffa59d8" containerID="35b893c825ea2cf30ddd270904b104304eed3db2ef55461fe4ab5b3edeb5c804" exitCode=0 Feb 25 13:37:55 crc kubenswrapper[4815]: I0225 13:37:55.928014 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" event={"ID":"3199ac65-2fd7-482f-b2fa-b5305ffa59d8","Type":"ContainerDied","Data":"35b893c825ea2cf30ddd270904b104304eed3db2ef55461fe4ab5b3edeb5c804"} Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.131580 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533778-xbfls"] Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.133191 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533778-xbfls" Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.135010 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.135185 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.135909 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.139459 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533778-xbfls"] Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.167887 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmzmb\" (UniqueName: \"kubernetes.io/projected/5139ef47-6709-46f0-8f06-46cadbbb36b1-kube-api-access-lmzmb\") pod \"auto-csr-approver-29533778-xbfls\" (UID: \"5139ef47-6709-46f0-8f06-46cadbbb36b1\") " pod="openshift-infra/auto-csr-approver-29533778-xbfls" Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.269983 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmzmb\" (UniqueName: \"kubernetes.io/projected/5139ef47-6709-46f0-8f06-46cadbbb36b1-kube-api-access-lmzmb\") pod \"auto-csr-approver-29533778-xbfls\" (UID: \"5139ef47-6709-46f0-8f06-46cadbbb36b1\") " pod="openshift-infra/auto-csr-approver-29533778-xbfls" Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.289345 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmzmb\" (UniqueName: \"kubernetes.io/projected/5139ef47-6709-46f0-8f06-46cadbbb36b1-kube-api-access-lmzmb\") pod \"auto-csr-approver-29533778-xbfls\" (UID: \"5139ef47-6709-46f0-8f06-46cadbbb36b1\") " pod="openshift-infra/auto-csr-approver-29533778-xbfls" Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.466630 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533778-xbfls" Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.807080 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" podUID="3199ac65-2fd7-482f-b2fa-b5305ffa59d8" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: connect: connection refused" Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.941623 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.990959 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-fernet-keys\") pod \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.990999 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-credential-keys\") pod \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.991024 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-scripts\") pod \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.991068 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-combined-ca-bundle\") pod \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.991199 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw2v6\" (UniqueName: \"kubernetes.io/projected/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-kube-api-access-rw2v6\") pod \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.991274 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-config-data\") pod \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\" (UID: \"c36a30b8-6a56-4f23-8c5e-7556199b2ad4\") " Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.997067 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-kube-api-access-rw2v6" (OuterVolumeSpecName: "kube-api-access-rw2v6") pod "c36a30b8-6a56-4f23-8c5e-7556199b2ad4" (UID: "c36a30b8-6a56-4f23-8c5e-7556199b2ad4"). InnerVolumeSpecName "kube-api-access-rw2v6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:00 crc kubenswrapper[4815]: I0225 13:38:00.997777 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c36a30b8-6a56-4f23-8c5e-7556199b2ad4" (UID: "c36a30b8-6a56-4f23-8c5e-7556199b2ad4"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:01 crc kubenswrapper[4815]: I0225 13:38:01.003914 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qs8h9" event={"ID":"c36a30b8-6a56-4f23-8c5e-7556199b2ad4","Type":"ContainerDied","Data":"d2218c3e8b1f23653d1a708ab1f4275aa530220b625011e8908d14cf8b45bec8"} Feb 25 13:38:01 crc kubenswrapper[4815]: I0225 13:38:01.004104 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2218c3e8b1f23653d1a708ab1f4275aa530220b625011e8908d14cf8b45bec8" Feb 25 13:38:01 crc kubenswrapper[4815]: I0225 13:38:01.003993 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qs8h9" Feb 25 13:38:01 crc kubenswrapper[4815]: I0225 13:38:01.013668 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c36a30b8-6a56-4f23-8c5e-7556199b2ad4" (UID: "c36a30b8-6a56-4f23-8c5e-7556199b2ad4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:01 crc kubenswrapper[4815]: I0225 13:38:01.014891 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-scripts" (OuterVolumeSpecName: "scripts") pod "c36a30b8-6a56-4f23-8c5e-7556199b2ad4" (UID: "c36a30b8-6a56-4f23-8c5e-7556199b2ad4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:01 crc kubenswrapper[4815]: I0225 13:38:01.024757 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c36a30b8-6a56-4f23-8c5e-7556199b2ad4" (UID: "c36a30b8-6a56-4f23-8c5e-7556199b2ad4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:01 crc kubenswrapper[4815]: I0225 13:38:01.027400 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-config-data" (OuterVolumeSpecName: "config-data") pod "c36a30b8-6a56-4f23-8c5e-7556199b2ad4" (UID: "c36a30b8-6a56-4f23-8c5e-7556199b2ad4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:01 crc kubenswrapper[4815]: I0225 13:38:01.095264 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw2v6\" (UniqueName: \"kubernetes.io/projected/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-kube-api-access-rw2v6\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:01 crc kubenswrapper[4815]: I0225 13:38:01.095453 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:01 crc kubenswrapper[4815]: I0225 13:38:01.095573 4815 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:01 crc kubenswrapper[4815]: I0225 13:38:01.095644 4815 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:01 crc kubenswrapper[4815]: I0225 13:38:01.095700 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:01 crc kubenswrapper[4815]: I0225 13:38:01.095757 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c36a30b8-6a56-4f23-8c5e-7556199b2ad4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.129146 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-qs8h9"] Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.140129 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-qs8h9"] Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.226016 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-4x94p"] Feb 25 13:38:02 crc kubenswrapper[4815]: E0225 13:38:02.226532 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c36a30b8-6a56-4f23-8c5e-7556199b2ad4" containerName="keystone-bootstrap" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.226554 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="c36a30b8-6a56-4f23-8c5e-7556199b2ad4" containerName="keystone-bootstrap" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.227443 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="c36a30b8-6a56-4f23-8c5e-7556199b2ad4" containerName="keystone-bootstrap" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.228151 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.230797 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.231804 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.231851 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-tftcl" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.231900 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.234094 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.234197 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4x94p"] Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.321679 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-fernet-keys\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.321743 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-scripts\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.321764 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-combined-ca-bundle\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.321848 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrpwb\" (UniqueName: \"kubernetes.io/projected/63427264-5351-40c4-aeff-d35ea829ccd4-kube-api-access-wrpwb\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.321921 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-credential-keys\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.321995 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-config-data\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.423595 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrpwb\" (UniqueName: \"kubernetes.io/projected/63427264-5351-40c4-aeff-d35ea829ccd4-kube-api-access-wrpwb\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.423679 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-credential-keys\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.423750 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-config-data\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.423826 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-fernet-keys\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.423854 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-scripts\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.423874 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-combined-ca-bundle\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.428070 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-credential-keys\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.428269 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-combined-ca-bundle\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.428578 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-fernet-keys\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.429154 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-config-data\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.430043 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-scripts\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.448826 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrpwb\" (UniqueName: \"kubernetes.io/projected/63427264-5351-40c4-aeff-d35ea829ccd4-kube-api-access-wrpwb\") pod \"keystone-bootstrap-4x94p\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.550386 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:02 crc kubenswrapper[4815]: E0225 13:38:02.622565 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 25 13:38:02 crc kubenswrapper[4815]: E0225 13:38:02.622954 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n95h688h584h699hd6h558hb5h9bhd6h5d6h5dh674h695h55dh55dhc7h676h77h67dh578h4h58dh594h57fh85h5f6h58h64ch68h585h548hfbq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mhhln,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-85878dd5b5-q9gfx_openstack(755e6fea-b2e0-4dfb-99e9-6ecad78e754d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 13:38:02 crc kubenswrapper[4815]: E0225 13:38:02.628606 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 25 13:38:02 crc kubenswrapper[4815]: E0225 13:38:02.628742 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nf5h5chbfh64chf6h567h64dh59h9h5bch589h647hb6hb4h65bh598hf8h9dh55h647h589hf7hb7h9fh5dfh9ch59dh667h565h77h69h5dbq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2xhv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5ff9b486b9-79sqs_openstack(ef5ee7f2-aefa-4107-8fa9-57902643ab8e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 13:38:02 crc kubenswrapper[4815]: E0225 13:38:02.628848 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-85878dd5b5-q9gfx" podUID="755e6fea-b2e0-4dfb-99e9-6ecad78e754d" Feb 25 13:38:02 crc kubenswrapper[4815]: E0225 13:38:02.632761 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5ff9b486b9-79sqs" podUID="ef5ee7f2-aefa-4107-8fa9-57902643ab8e" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.693146 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.730480 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-combined-ca-bundle\") pod \"26e0f670-91b0-4bed-8a45-596bcf916ce3\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.730594 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-internal-tls-certs\") pod \"26e0f670-91b0-4bed-8a45-596bcf916ce3\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.730690 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"26e0f670-91b0-4bed-8a45-596bcf916ce3\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.730747 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26e0f670-91b0-4bed-8a45-596bcf916ce3-logs\") pod \"26e0f670-91b0-4bed-8a45-596bcf916ce3\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.730832 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-scripts\") pod \"26e0f670-91b0-4bed-8a45-596bcf916ce3\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.730872 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-config-data\") pod \"26e0f670-91b0-4bed-8a45-596bcf916ce3\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.731996 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2t7qn\" (UniqueName: \"kubernetes.io/projected/26e0f670-91b0-4bed-8a45-596bcf916ce3-kube-api-access-2t7qn\") pod \"26e0f670-91b0-4bed-8a45-596bcf916ce3\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.732136 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26e0f670-91b0-4bed-8a45-596bcf916ce3-httpd-run\") pod \"26e0f670-91b0-4bed-8a45-596bcf916ce3\" (UID: \"26e0f670-91b0-4bed-8a45-596bcf916ce3\") " Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.733451 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26e0f670-91b0-4bed-8a45-596bcf916ce3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "26e0f670-91b0-4bed-8a45-596bcf916ce3" (UID: "26e0f670-91b0-4bed-8a45-596bcf916ce3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.736439 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26e0f670-91b0-4bed-8a45-596bcf916ce3-logs" (OuterVolumeSpecName: "logs") pod "26e0f670-91b0-4bed-8a45-596bcf916ce3" (UID: "26e0f670-91b0-4bed-8a45-596bcf916ce3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.752723 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26e0f670-91b0-4bed-8a45-596bcf916ce3-kube-api-access-2t7qn" (OuterVolumeSpecName: "kube-api-access-2t7qn") pod "26e0f670-91b0-4bed-8a45-596bcf916ce3" (UID: "26e0f670-91b0-4bed-8a45-596bcf916ce3"). InnerVolumeSpecName "kube-api-access-2t7qn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.752784 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "26e0f670-91b0-4bed-8a45-596bcf916ce3" (UID: "26e0f670-91b0-4bed-8a45-596bcf916ce3"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.753207 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-scripts" (OuterVolumeSpecName: "scripts") pod "26e0f670-91b0-4bed-8a45-596bcf916ce3" (UID: "26e0f670-91b0-4bed-8a45-596bcf916ce3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.762689 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26e0f670-91b0-4bed-8a45-596bcf916ce3" (UID: "26e0f670-91b0-4bed-8a45-596bcf916ce3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.787740 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "26e0f670-91b0-4bed-8a45-596bcf916ce3" (UID: "26e0f670-91b0-4bed-8a45-596bcf916ce3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.804541 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-config-data" (OuterVolumeSpecName: "config-data") pod "26e0f670-91b0-4bed-8a45-596bcf916ce3" (UID: "26e0f670-91b0-4bed-8a45-596bcf916ce3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.835034 4815 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.835084 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26e0f670-91b0-4bed-8a45-596bcf916ce3-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.835097 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.835117 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.835132 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2t7qn\" (UniqueName: \"kubernetes.io/projected/26e0f670-91b0-4bed-8a45-596bcf916ce3-kube-api-access-2t7qn\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.835146 4815 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26e0f670-91b0-4bed-8a45-596bcf916ce3-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.835155 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.835166 4815 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/26e0f670-91b0-4bed-8a45-596bcf916ce3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.865244 4815 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.936329 4815 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:02 crc kubenswrapper[4815]: I0225 13:38:02.949315 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c36a30b8-6a56-4f23-8c5e-7556199b2ad4" path="/var/lib/kubelet/pods/c36a30b8-6a56-4f23-8c5e-7556199b2ad4/volumes" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.020103 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"26e0f670-91b0-4bed-8a45-596bcf916ce3","Type":"ContainerDied","Data":"28f8e62b0377f288e99cdd111648bbc41d158ac25445d742700df51f6cdb097a"} Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.020243 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.118700 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.137563 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.158031 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 13:38:03 crc kubenswrapper[4815]: E0225 13:38:03.158432 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26e0f670-91b0-4bed-8a45-596bcf916ce3" containerName="glance-log" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.158444 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="26e0f670-91b0-4bed-8a45-596bcf916ce3" containerName="glance-log" Feb 25 13:38:03 crc kubenswrapper[4815]: E0225 13:38:03.158461 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26e0f670-91b0-4bed-8a45-596bcf916ce3" containerName="glance-httpd" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.158468 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="26e0f670-91b0-4bed-8a45-596bcf916ce3" containerName="glance-httpd" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.158656 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="26e0f670-91b0-4bed-8a45-596bcf916ce3" containerName="glance-log" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.158674 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="26e0f670-91b0-4bed-8a45-596bcf916ce3" containerName="glance-httpd" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.159519 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.159598 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.169058 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.169312 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.245423 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-config-data\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.245464 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gshpz\" (UniqueName: \"kubernetes.io/projected/532b0818-f59f-4b9d-a4fd-6a976bcdb886-kube-api-access-gshpz\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.245498 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.245533 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.245584 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/532b0818-f59f-4b9d-a4fd-6a976bcdb886-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.245619 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-scripts\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.245666 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.245680 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/532b0818-f59f-4b9d-a4fd-6a976bcdb886-logs\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.349025 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/532b0818-f59f-4b9d-a4fd-6a976bcdb886-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.349090 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-scripts\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.349146 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.349164 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/532b0818-f59f-4b9d-a4fd-6a976bcdb886-logs\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.349230 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-config-data\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.349246 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gshpz\" (UniqueName: \"kubernetes.io/projected/532b0818-f59f-4b9d-a4fd-6a976bcdb886-kube-api-access-gshpz\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.349266 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.349288 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.349684 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.350436 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/532b0818-f59f-4b9d-a4fd-6a976bcdb886-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.350489 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/532b0818-f59f-4b9d-a4fd-6a976bcdb886-logs\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.353883 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-config-data\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.357585 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.358219 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.358833 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-scripts\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.368104 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gshpz\" (UniqueName: \"kubernetes.io/projected/532b0818-f59f-4b9d-a4fd-6a976bcdb886-kube-api-access-gshpz\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.407002 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:38:03 crc kubenswrapper[4815]: I0225 13:38:03.501358 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 13:38:04 crc kubenswrapper[4815]: I0225 13:38:04.946683 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26e0f670-91b0-4bed-8a45-596bcf916ce3" path="/var/lib/kubelet/pods/26e0f670-91b0-4bed-8a45-596bcf916ce3/volumes" Feb 25 13:38:05 crc kubenswrapper[4815]: I0225 13:38:05.042827 4815 generic.go:334] "Generic (PLEG): container finished" podID="ab03b49b-be69-4e1d-99e7-2f71a602b110" containerID="4dd4c4d4dba488e7472dcc7279f6c3ad41d76de7b59eb5521533e41a8e0c8a40" exitCode=0 Feb 25 13:38:05 crc kubenswrapper[4815]: I0225 13:38:05.042884 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2kr57" event={"ID":"ab03b49b-be69-4e1d-99e7-2f71a602b110","Type":"ContainerDied","Data":"4dd4c4d4dba488e7472dcc7279f6c3ad41d76de7b59eb5521533e41a8e0c8a40"} Feb 25 13:38:10 crc kubenswrapper[4815]: I0225 13:38:10.807740 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" podUID="3199ac65-2fd7-482f-b2fa-b5305ffa59d8" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: i/o timeout" Feb 25 13:38:10 crc kubenswrapper[4815]: I0225 13:38:10.808696 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:38:11 crc kubenswrapper[4815]: E0225 13:38:11.337846 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Feb 25 13:38:11 crc kubenswrapper[4815]: E0225 13:38:11.337991 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7spnb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-d4ssg_openstack(7bdc9071-e863-4108-9234-4f201732f435): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 13:38:11 crc kubenswrapper[4815]: E0225 13:38:11.339146 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-d4ssg" podUID="7bdc9071-e863-4108-9234-4f201732f435" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.373339 4815 scope.go:117] "RemoveContainer" containerID="d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5" Feb 25 13:38:11 crc kubenswrapper[4815]: E0225 13:38:11.373917 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5\": container with ID starting with d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5 not found: ID does not exist" containerID="d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.374001 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5"} err="failed to get container status \"d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5\": rpc error: code = NotFound desc = could not find container \"d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5\": container with ID starting with d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5 not found: ID does not exist" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.374036 4815 scope.go:117] "RemoveContainer" containerID="fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89" Feb 25 13:38:11 crc kubenswrapper[4815]: E0225 13:38:11.374377 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89\": container with ID starting with fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89 not found: ID does not exist" containerID="fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.374411 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89"} err="failed to get container status \"fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89\": rpc error: code = NotFound desc = could not find container \"fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89\": container with ID starting with fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89 not found: ID does not exist" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.374433 4815 scope.go:117] "RemoveContainer" containerID="d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.374726 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5"} err="failed to get container status \"d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5\": rpc error: code = NotFound desc = could not find container \"d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5\": container with ID starting with d2a46a2abd8ef6e198140dcba106e7502b47021695ffc492190f2a9cc74805d5 not found: ID does not exist" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.374766 4815 scope.go:117] "RemoveContainer" containerID="fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.375038 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89"} err="failed to get container status \"fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89\": rpc error: code = NotFound desc = could not find container \"fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89\": container with ID starting with fa25b5bb26a002a069794bbc2670c84cfe197c5326a6f2668b38e5c4db703e89 not found: ID does not exist" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.375057 4815 scope.go:117] "RemoveContainer" containerID="7df816bfd022b3c0e86bc9e077e3a98903d2b2854df2d805581f0efa43ba685f" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.509992 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2kr57" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.518822 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.525652 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.543007 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723228 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2xhv\" (UniqueName: \"kubernetes.io/projected/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-kube-api-access-s2xhv\") pod \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723285 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab03b49b-be69-4e1d-99e7-2f71a602b110-combined-ca-bundle\") pod \"ab03b49b-be69-4e1d-99e7-2f71a602b110\" (UID: \"ab03b49b-be69-4e1d-99e7-2f71a602b110\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723319 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-logs\") pod \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723372 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-horizon-secret-key\") pod \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723434 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngwbf\" (UniqueName: \"kubernetes.io/projected/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-kube-api-access-ngwbf\") pod \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723481 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-ovsdbserver-nb\") pod \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723525 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-config\") pod \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723563 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhhln\" (UniqueName: \"kubernetes.io/projected/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-kube-api-access-mhhln\") pod \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723586 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-dns-swift-storage-0\") pod \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723614 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-config-data\") pod \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723629 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-scripts\") pod \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723648 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-dns-svc\") pod \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723680 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9v8w7\" (UniqueName: \"kubernetes.io/projected/ab03b49b-be69-4e1d-99e7-2f71a602b110-kube-api-access-9v8w7\") pod \"ab03b49b-be69-4e1d-99e7-2f71a602b110\" (UID: \"ab03b49b-be69-4e1d-99e7-2f71a602b110\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723709 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-scripts\") pod \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723729 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-horizon-secret-key\") pod \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\" (UID: \"755e6fea-b2e0-4dfb-99e9-6ecad78e754d\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723745 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-config-data\") pod \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723764 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ab03b49b-be69-4e1d-99e7-2f71a602b110-config\") pod \"ab03b49b-be69-4e1d-99e7-2f71a602b110\" (UID: \"ab03b49b-be69-4e1d-99e7-2f71a602b110\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723817 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-ovsdbserver-sb\") pod \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\" (UID: \"3199ac65-2fd7-482f-b2fa-b5305ffa59d8\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723873 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-logs\") pod \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\" (UID: \"ef5ee7f2-aefa-4107-8fa9-57902643ab8e\") " Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.723697 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-logs" (OuterVolumeSpecName: "logs") pod "755e6fea-b2e0-4dfb-99e9-6ecad78e754d" (UID: "755e6fea-b2e0-4dfb-99e9-6ecad78e754d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.724526 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-scripts" (OuterVolumeSpecName: "scripts") pod "755e6fea-b2e0-4dfb-99e9-6ecad78e754d" (UID: "755e6fea-b2e0-4dfb-99e9-6ecad78e754d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.724875 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-config-data" (OuterVolumeSpecName: "config-data") pod "755e6fea-b2e0-4dfb-99e9-6ecad78e754d" (UID: "755e6fea-b2e0-4dfb-99e9-6ecad78e754d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.725244 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-logs" (OuterVolumeSpecName: "logs") pod "ef5ee7f2-aefa-4107-8fa9-57902643ab8e" (UID: "ef5ee7f2-aefa-4107-8fa9-57902643ab8e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.725455 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-scripts" (OuterVolumeSpecName: "scripts") pod "ef5ee7f2-aefa-4107-8fa9-57902643ab8e" (UID: "ef5ee7f2-aefa-4107-8fa9-57902643ab8e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.727458 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-config-data" (OuterVolumeSpecName: "config-data") pod "ef5ee7f2-aefa-4107-8fa9-57902643ab8e" (UID: "ef5ee7f2-aefa-4107-8fa9-57902643ab8e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.728476 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ef5ee7f2-aefa-4107-8fa9-57902643ab8e" (UID: "ef5ee7f2-aefa-4107-8fa9-57902643ab8e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.728526 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-kube-api-access-s2xhv" (OuterVolumeSpecName: "kube-api-access-s2xhv") pod "ef5ee7f2-aefa-4107-8fa9-57902643ab8e" (UID: "ef5ee7f2-aefa-4107-8fa9-57902643ab8e"). InnerVolumeSpecName "kube-api-access-s2xhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.728471 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-kube-api-access-mhhln" (OuterVolumeSpecName: "kube-api-access-mhhln") pod "755e6fea-b2e0-4dfb-99e9-6ecad78e754d" (UID: "755e6fea-b2e0-4dfb-99e9-6ecad78e754d"). InnerVolumeSpecName "kube-api-access-mhhln". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.730200 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-kube-api-access-ngwbf" (OuterVolumeSpecName: "kube-api-access-ngwbf") pod "3199ac65-2fd7-482f-b2fa-b5305ffa59d8" (UID: "3199ac65-2fd7-482f-b2fa-b5305ffa59d8"). InnerVolumeSpecName "kube-api-access-ngwbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.731144 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "755e6fea-b2e0-4dfb-99e9-6ecad78e754d" (UID: "755e6fea-b2e0-4dfb-99e9-6ecad78e754d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.731894 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab03b49b-be69-4e1d-99e7-2f71a602b110-kube-api-access-9v8w7" (OuterVolumeSpecName: "kube-api-access-9v8w7") pod "ab03b49b-be69-4e1d-99e7-2f71a602b110" (UID: "ab03b49b-be69-4e1d-99e7-2f71a602b110"). InnerVolumeSpecName "kube-api-access-9v8w7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.748819 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab03b49b-be69-4e1d-99e7-2f71a602b110-config" (OuterVolumeSpecName: "config") pod "ab03b49b-be69-4e1d-99e7-2f71a602b110" (UID: "ab03b49b-be69-4e1d-99e7-2f71a602b110"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.751240 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab03b49b-be69-4e1d-99e7-2f71a602b110-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab03b49b-be69-4e1d-99e7-2f71a602b110" (UID: "ab03b49b-be69-4e1d-99e7-2f71a602b110"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.765626 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3199ac65-2fd7-482f-b2fa-b5305ffa59d8" (UID: "3199ac65-2fd7-482f-b2fa-b5305ffa59d8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.768464 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-config" (OuterVolumeSpecName: "config") pod "3199ac65-2fd7-482f-b2fa-b5305ffa59d8" (UID: "3199ac65-2fd7-482f-b2fa-b5305ffa59d8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.769756 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3199ac65-2fd7-482f-b2fa-b5305ffa59d8" (UID: "3199ac65-2fd7-482f-b2fa-b5305ffa59d8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.781492 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3199ac65-2fd7-482f-b2fa-b5305ffa59d8" (UID: "3199ac65-2fd7-482f-b2fa-b5305ffa59d8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.788585 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3199ac65-2fd7-482f-b2fa-b5305ffa59d8" (UID: "3199ac65-2fd7-482f-b2fa-b5305ffa59d8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826190 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826259 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2xhv\" (UniqueName: \"kubernetes.io/projected/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-kube-api-access-s2xhv\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826273 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab03b49b-be69-4e1d-99e7-2f71a602b110-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826282 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826291 4815 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826300 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngwbf\" (UniqueName: \"kubernetes.io/projected/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-kube-api-access-ngwbf\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826308 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826317 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826324 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhhln\" (UniqueName: \"kubernetes.io/projected/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-kube-api-access-mhhln\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826332 4815 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826340 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826350 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826357 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826365 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9v8w7\" (UniqueName: \"kubernetes.io/projected/ab03b49b-be69-4e1d-99e7-2f71a602b110-kube-api-access-9v8w7\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826372 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826379 4815 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/755e6fea-b2e0-4dfb-99e9-6ecad78e754d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826387 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef5ee7f2-aefa-4107-8fa9-57902643ab8e-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826394 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ab03b49b-be69-4e1d-99e7-2f71a602b110-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:11 crc kubenswrapper[4815]: I0225 13:38:11.826401 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3199ac65-2fd7-482f-b2fa-b5305ffa59d8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.104715 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2kr57" event={"ID":"ab03b49b-be69-4e1d-99e7-2f71a602b110","Type":"ContainerDied","Data":"b2208242ae9be094b621e23d25975615ccd3e0c5acf428274eb4238accb00bc8"} Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.105055 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2208242ae9be094b621e23d25975615ccd3e0c5acf428274eb4238accb00bc8" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.105155 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2kr57" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.111232 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" event={"ID":"3199ac65-2fd7-482f-b2fa-b5305ffa59d8","Type":"ContainerDied","Data":"a0aa0fb524582fe979521956cdf24a03caa513bc0d1459cb333dc11d2ea37f61"} Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.111292 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.117006 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85878dd5b5-q9gfx" event={"ID":"755e6fea-b2e0-4dfb-99e9-6ecad78e754d","Type":"ContainerDied","Data":"f09c98a586f8643d63288327c5a64153cd634fda881aac812423d2b992ea46c9"} Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.117085 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85878dd5b5-q9gfx" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.118742 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5ff9b486b9-79sqs" event={"ID":"ef5ee7f2-aefa-4107-8fa9-57902643ab8e","Type":"ContainerDied","Data":"6707fce2a9afc6ed1a5c272214047fa2a2849dbc68b46d6f84b3f659ee0652ce"} Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.118847 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5ff9b486b9-79sqs" Feb 25 13:38:12 crc kubenswrapper[4815]: E0225 13:38:12.128030 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-d4ssg" podUID="7bdc9071-e863-4108-9234-4f201732f435" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.173391 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-5lnz6"] Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.180441 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-5lnz6"] Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.228261 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85878dd5b5-q9gfx"] Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.236684 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-85878dd5b5-q9gfx"] Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.251779 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5ff9b486b9-79sqs"] Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.258223 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5ff9b486b9-79sqs"] Feb 25 13:38:12 crc kubenswrapper[4815]: E0225 13:38:12.667562 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Feb 25 13:38:12 crc kubenswrapper[4815]: E0225 13:38:12.667745 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j9dbn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-576gs_openstack(04e3d721-2179-4457-8cc6-72af3b3f9257): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 13:38:12 crc kubenswrapper[4815]: E0225 13:38:12.668966 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-576gs" podUID="04e3d721-2179-4457-8cc6-72af3b3f9257" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.679637 4815 scope.go:117] "RemoveContainer" containerID="c6a74f1c0e7af9d7ae823c955b720f65bcea3698261db790707090128b3f9024" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.804375 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-gggp2"] Feb 25 13:38:12 crc kubenswrapper[4815]: E0225 13:38:12.804878 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab03b49b-be69-4e1d-99e7-2f71a602b110" containerName="neutron-db-sync" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.804914 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab03b49b-be69-4e1d-99e7-2f71a602b110" containerName="neutron-db-sync" Feb 25 13:38:12 crc kubenswrapper[4815]: E0225 13:38:12.804954 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3199ac65-2fd7-482f-b2fa-b5305ffa59d8" containerName="dnsmasq-dns" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.804962 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="3199ac65-2fd7-482f-b2fa-b5305ffa59d8" containerName="dnsmasq-dns" Feb 25 13:38:12 crc kubenswrapper[4815]: E0225 13:38:12.804981 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3199ac65-2fd7-482f-b2fa-b5305ffa59d8" containerName="init" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.804989 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="3199ac65-2fd7-482f-b2fa-b5305ffa59d8" containerName="init" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.805223 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="3199ac65-2fd7-482f-b2fa-b5305ffa59d8" containerName="dnsmasq-dns" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.805238 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab03b49b-be69-4e1d-99e7-2f71a602b110" containerName="neutron-db-sync" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.806380 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.832382 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-gggp2"] Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.857046 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44jd4\" (UniqueName: \"kubernetes.io/projected/ea8d6d97-73b0-4db9-8712-31eb46ef0105-kube-api-access-44jd4\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.857096 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.857129 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.857227 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.857312 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.857354 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-config\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.889140 4815 scope.go:117] "RemoveContainer" containerID="35b893c825ea2cf30ddd270904b104304eed3db2ef55461fe4ab5b3edeb5c804" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.910597 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6fdb68948-zzdwv"] Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.914547 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.916887 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.916896 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qmrd2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.917114 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.917337 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.924226 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6fdb68948-zzdwv"] Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.958768 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.958852 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.958893 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-config\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.958991 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44jd4\" (UniqueName: \"kubernetes.io/projected/ea8d6d97-73b0-4db9-8712-31eb46ef0105-kube-api-access-44jd4\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.959021 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.959051 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.960617 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.961206 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-config\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.961952 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.961976 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.965552 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3199ac65-2fd7-482f-b2fa-b5305ffa59d8" path="/var/lib/kubelet/pods/3199ac65-2fd7-482f-b2fa-b5305ffa59d8/volumes" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.966602 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="755e6fea-b2e0-4dfb-99e9-6ecad78e754d" path="/var/lib/kubelet/pods/755e6fea-b2e0-4dfb-99e9-6ecad78e754d/volumes" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.967176 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.971246 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef5ee7f2-aefa-4107-8fa9-57902643ab8e" path="/var/lib/kubelet/pods/ef5ee7f2-aefa-4107-8fa9-57902643ab8e/volumes" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.986374 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44jd4\" (UniqueName: \"kubernetes.io/projected/ea8d6d97-73b0-4db9-8712-31eb46ef0105-kube-api-access-44jd4\") pod \"dnsmasq-dns-5ccc5c4795-gggp2\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:12 crc kubenswrapper[4815]: I0225 13:38:12.992753 4815 scope.go:117] "RemoveContainer" containerID="4706cd5360f951abc9d3faa9a0e981097dc091f1e28467b705fee874282528dc" Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.060942 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-config\") pod \"neutron-6fdb68948-zzdwv\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.060985 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-httpd-config\") pod \"neutron-6fdb68948-zzdwv\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.061029 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxsqh\" (UniqueName: \"kubernetes.io/projected/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-kube-api-access-zxsqh\") pod \"neutron-6fdb68948-zzdwv\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.061078 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-combined-ca-bundle\") pod \"neutron-6fdb68948-zzdwv\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.061103 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-ovndb-tls-certs\") pod \"neutron-6fdb68948-zzdwv\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:13 crc kubenswrapper[4815]: E0225 13:38:13.150396 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-576gs" podUID="04e3d721-2179-4457-8cc6-72af3b3f9257" Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.205595 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.206256 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-httpd-config\") pod \"neutron-6fdb68948-zzdwv\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.206327 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxsqh\" (UniqueName: \"kubernetes.io/projected/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-kube-api-access-zxsqh\") pod \"neutron-6fdb68948-zzdwv\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.206370 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-combined-ca-bundle\") pod \"neutron-6fdb68948-zzdwv\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.206873 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-ovndb-tls-certs\") pod \"neutron-6fdb68948-zzdwv\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.207359 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-config\") pod \"neutron-6fdb68948-zzdwv\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.211099 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-ovndb-tls-certs\") pod \"neutron-6fdb68948-zzdwv\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.211480 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-httpd-config\") pod \"neutron-6fdb68948-zzdwv\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.211533 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-config\") pod \"neutron-6fdb68948-zzdwv\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.213166 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-combined-ca-bundle\") pod \"neutron-6fdb68948-zzdwv\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.228349 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxsqh\" (UniqueName: \"kubernetes.io/projected/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-kube-api-access-zxsqh\") pod \"neutron-6fdb68948-zzdwv\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.304120 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7657655688-k9k6t"] Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.308879 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:13 crc kubenswrapper[4815]: W0225 13:38:13.364120 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf10b8779_9d5e_427d_847c_3250816e89b9.slice/crio-82388074bed9c097d092cdeac1b178292f8a1404ce2a51d86d46ff24721ef729 WatchSource:0}: Error finding container 82388074bed9c097d092cdeac1b178292f8a1404ce2a51d86d46ff24721ef729: Status 404 returned error can't find the container with id 82388074bed9c097d092cdeac1b178292f8a1404ce2a51d86d46ff24721ef729 Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.631841 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4x94p"] Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.667409 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6497d456d-s5g59"] Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.733530 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533778-xbfls"] Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.756702 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-gggp2"] Feb 25 13:38:13 crc kubenswrapper[4815]: I0225 13:38:13.889932 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.012230 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6fdb68948-zzdwv"] Feb 25 13:38:14 crc kubenswrapper[4815]: W0225 13:38:14.034658 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb02ef5af_64a9_4fbc_b312_30ac0ffab7ee.slice/crio-cff5a29db301d89f37170debc4530bf24f0e2220e4646766d268e8fca776c2ed WatchSource:0}: Error finding container cff5a29db301d89f37170debc4530bf24f0e2220e4646766d268e8fca776c2ed: Status 404 returned error can't find the container with id cff5a29db301d89f37170debc4530bf24f0e2220e4646766d268e8fca776c2ed Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.173743 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6497d456d-s5g59" event={"ID":"5d076fbb-4499-456f-82ee-211a8909fc99","Type":"ContainerStarted","Data":"37b004a1e59b2810a4a4e2a9e741191118c1925114ad66b49a3a6e92a2d9e11e"} Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.173791 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6497d456d-s5g59" event={"ID":"5d076fbb-4499-456f-82ee-211a8909fc99","Type":"ContainerStarted","Data":"79c17dbf7bf2ac9651809fd2e4d5f718ceebe6086b34ddd6ad5c389628898018"} Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.177883 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533778-xbfls" event={"ID":"5139ef47-6709-46f0-8f06-46cadbbb36b1","Type":"ContainerStarted","Data":"c74ec21c6d9c75a615e021af549d77630e6c23e5691db606f80ff9d7d94ac7d5"} Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.182276 4815 generic.go:334] "Generic (PLEG): container finished" podID="ea8d6d97-73b0-4db9-8712-31eb46ef0105" containerID="cef0e7e70b9abdc983c0fa359e7d139f9e6011b9efe129abfd2d41f709ac71aa" exitCode=0 Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.182339 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" event={"ID":"ea8d6d97-73b0-4db9-8712-31eb46ef0105","Type":"ContainerDied","Data":"cef0e7e70b9abdc983c0fa359e7d139f9e6011b9efe129abfd2d41f709ac71aa"} Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.182371 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" event={"ID":"ea8d6d97-73b0-4db9-8712-31eb46ef0105","Type":"ContainerStarted","Data":"612731f99c90e554c44c1b407cdc68e3b8d6131a290dcd415a0a14c2859bcac9"} Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.191783 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c455f8ff5-tkdz5" event={"ID":"d9c18809-9f62-4b7e-8103-df382e9b4fa6","Type":"ContainerStarted","Data":"3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f"} Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.191827 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c455f8ff5-tkdz5" event={"ID":"d9c18809-9f62-4b7e-8103-df382e9b4fa6","Type":"ContainerStarted","Data":"46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102"} Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.191869 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c455f8ff5-tkdz5" podUID="d9c18809-9f62-4b7e-8103-df382e9b4fa6" containerName="horizon-log" containerID="cri-o://46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102" gracePeriod=30 Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.191900 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c455f8ff5-tkdz5" podUID="d9c18809-9f62-4b7e-8103-df382e9b4fa6" containerName="horizon" containerID="cri-o://3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f" gracePeriod=30 Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.235543 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-c455f8ff5-tkdz5" podStartSLOduration=5.125907484 podStartE2EDuration="31.235523056s" podCreationTimestamp="2026-02-25 13:37:43 +0000 UTC" firstStartedPulling="2026-02-25 13:37:45.267403609 +0000 UTC m=+1023.068501663" lastFinishedPulling="2026-02-25 13:38:11.377019181 +0000 UTC m=+1049.178117235" observedRunningTime="2026-02-25 13:38:14.230937593 +0000 UTC m=+1052.032035657" watchObservedRunningTime="2026-02-25 13:38:14.235523056 +0000 UTC m=+1052.036621120" Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.236134 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3643568-6595-4c73-9c83-be6e28fabe83","Type":"ContainerStarted","Data":"dc7434d19b0960010335e00c9621ca0164b153f5cd542c93cb16504e4883f085"} Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.240374 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4x94p" event={"ID":"63427264-5351-40c4-aeff-d35ea829ccd4","Type":"ContainerStarted","Data":"aa83b80cdebfae2c0db1d74c7475c986c41dd26e06d53c3e44db3c6c664fa6df"} Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.240426 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4x94p" event={"ID":"63427264-5351-40c4-aeff-d35ea829ccd4","Type":"ContainerStarted","Data":"65120763d9188de1d47032dd07435216a8381d897169e3f1e152f2bf2ce920e7"} Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.243018 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mgztm" event={"ID":"ee89a5e7-e5a4-412f-a52a-511dfca61265","Type":"ContainerStarted","Data":"c5a9af64e2d4e57b092b43e4a47fe7a5daffc4c0026054c3e700d2da1bb3823d"} Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.246852 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"532b0818-f59f-4b9d-a4fd-6a976bcdb886","Type":"ContainerStarted","Data":"17b60340b74377f53425b3acfc7f303c43fcbbba844febba81bcb8cf2bc6485e"} Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.252029 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7657655688-k9k6t" event={"ID":"f10b8779-9d5e-427d-847c-3250816e89b9","Type":"ContainerStarted","Data":"75501fc87635307532d133cff48e3928bac0219aedc3fd85b028ea8dd2144008"} Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.252071 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7657655688-k9k6t" event={"ID":"f10b8779-9d5e-427d-847c-3250816e89b9","Type":"ContainerStarted","Data":"040d3ba0f42cccda404c6725e0ee9521a79d623d53d7858a6892550294afb0c7"} Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.252108 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7657655688-k9k6t" event={"ID":"f10b8779-9d5e-427d-847c-3250816e89b9","Type":"ContainerStarted","Data":"82388074bed9c097d092cdeac1b178292f8a1404ce2a51d86d46ff24721ef729"} Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.258856 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fdb68948-zzdwv" event={"ID":"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee","Type":"ContainerStarted","Data":"cff5a29db301d89f37170debc4530bf24f0e2220e4646766d268e8fca776c2ed"} Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.260255 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-4x94p" podStartSLOduration=12.260234689 podStartE2EDuration="12.260234689s" podCreationTimestamp="2026-02-25 13:38:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:14.255351026 +0000 UTC m=+1052.056449080" watchObservedRunningTime="2026-02-25 13:38:14.260234689 +0000 UTC m=+1052.061332743" Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.280816 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7657655688-k9k6t" podStartSLOduration=21.28080018 podStartE2EDuration="21.28080018s" podCreationTimestamp="2026-02-25 13:37:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:14.276912257 +0000 UTC m=+1052.078010311" watchObservedRunningTime="2026-02-25 13:38:14.28080018 +0000 UTC m=+1052.081898234" Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.296176 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-mgztm" podStartSLOduration=4.600492351 podStartE2EDuration="30.296151689s" podCreationTimestamp="2026-02-25 13:37:44 +0000 UTC" firstStartedPulling="2026-02-25 13:37:45.681305852 +0000 UTC m=+1023.482403896" lastFinishedPulling="2026-02-25 13:38:11.37696516 +0000 UTC m=+1049.178063234" observedRunningTime="2026-02-25 13:38:14.291362738 +0000 UTC m=+1052.092460792" watchObservedRunningTime="2026-02-25 13:38:14.296151689 +0000 UTC m=+1052.097249743" Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.516667 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:38:14 crc kubenswrapper[4815]: I0225 13:38:14.974918 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.295910 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fdb68948-zzdwv" event={"ID":"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee","Type":"ContainerStarted","Data":"e8b05302d38f26d33ea76e9b4625bfd62a5ec72ad4bcb7929790aa9bc15eab5d"} Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.295962 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fdb68948-zzdwv" event={"ID":"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee","Type":"ContainerStarted","Data":"ef9ce2c46a0c5e9b4cda1d459ecff2c694b65b93036e13dbb2d1c30c9858cde7"} Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.296015 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.315763 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6497d456d-s5g59" event={"ID":"5d076fbb-4499-456f-82ee-211a8909fc99","Type":"ContainerStarted","Data":"9fba6bb1e21ffc87dd876586c19b77148678fd408e930bd6088e1ed1c188d47d"} Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.321008 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6fdb68948-zzdwv" podStartSLOduration=3.320992085 podStartE2EDuration="3.320992085s" podCreationTimestamp="2026-02-25 13:38:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:15.320081299 +0000 UTC m=+1053.121179353" watchObservedRunningTime="2026-02-25 13:38:15.320992085 +0000 UTC m=+1053.122090139" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.328409 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533778-xbfls" event={"ID":"5139ef47-6709-46f0-8f06-46cadbbb36b1","Type":"ContainerStarted","Data":"d09a900a7289126fcf20dcc95e26d94934df340a2b7e2ea2c89358cd26e4b697"} Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.337719 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" event={"ID":"ea8d6d97-73b0-4db9-8712-31eb46ef0105","Type":"ContainerStarted","Data":"0c4fecacd29410071ab5edabaf794d418b4b3db1af96dc7c45b4fdc9554e3452"} Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.340311 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.351434 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"532b0818-f59f-4b9d-a4fd-6a976bcdb886","Type":"ContainerStarted","Data":"7b7091745b0e6c271d6f260f9e7b2c007d9f3629133f78a3ad3bfd00245e119b"} Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.352368 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-645c577cc7-vpgnt"] Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.356054 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.364572 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.364875 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.372700 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6497d456d-s5g59" podStartSLOduration=22.372680456 podStartE2EDuration="22.372680456s" podCreationTimestamp="2026-02-25 13:37:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:15.35265113 +0000 UTC m=+1053.153749184" watchObservedRunningTime="2026-02-25 13:38:15.372680456 +0000 UTC m=+1053.173778510" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.434828 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-645c577cc7-vpgnt"] Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.436018 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533778-xbfls" podStartSLOduration=14.476510521 podStartE2EDuration="15.436000527s" podCreationTimestamp="2026-02-25 13:38:00 +0000 UTC" firstStartedPulling="2026-02-25 13:38:13.769279048 +0000 UTC m=+1051.570377102" lastFinishedPulling="2026-02-25 13:38:14.728769054 +0000 UTC m=+1052.529867108" observedRunningTime="2026-02-25 13:38:15.371611305 +0000 UTC m=+1053.172709359" watchObservedRunningTime="2026-02-25 13:38:15.436000527 +0000 UTC m=+1053.237098581" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.445729 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" podStartSLOduration=3.4457145909999998 podStartE2EDuration="3.445714591s" podCreationTimestamp="2026-02-25 13:38:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:15.406551926 +0000 UTC m=+1053.207649980" watchObservedRunningTime="2026-02-25 13:38:15.445714591 +0000 UTC m=+1053.246812645" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.461019 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-ovndb-tls-certs\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.461142 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-internal-tls-certs\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.461259 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-httpd-config\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.461326 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-962nr\" (UniqueName: \"kubernetes.io/projected/145d5de3-c56d-4209-b3b7-348f2f6d2365-kube-api-access-962nr\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.461368 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-config\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.461382 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-combined-ca-bundle\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.461446 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-public-tls-certs\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.564031 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-internal-tls-certs\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.564597 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-httpd-config\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.564938 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-962nr\" (UniqueName: \"kubernetes.io/projected/145d5de3-c56d-4209-b3b7-348f2f6d2365-kube-api-access-962nr\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.565027 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-config\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.565057 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-combined-ca-bundle\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.565292 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-public-tls-certs\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.566078 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-ovndb-tls-certs\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.571430 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-config\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.571542 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-httpd-config\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.572142 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-ovndb-tls-certs\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.573230 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-internal-tls-certs\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.574670 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-combined-ca-bundle\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.575167 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-public-tls-certs\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.583310 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-962nr\" (UniqueName: \"kubernetes.io/projected/145d5de3-c56d-4209-b3b7-348f2f6d2365-kube-api-access-962nr\") pod \"neutron-645c577cc7-vpgnt\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.683993 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:15 crc kubenswrapper[4815]: W0225 13:38:15.732211 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49be5892_2d7d_4e75_bf93_6a1d16e769fe.slice/crio-d27c3e23771676c3511779299bde6302a854d8941f0e8383263c10c3b0255c63 WatchSource:0}: Error finding container d27c3e23771676c3511779299bde6302a854d8941f0e8383263c10c3b0255c63: Status 404 returned error can't find the container with id d27c3e23771676c3511779299bde6302a854d8941f0e8383263c10c3b0255c63 Feb 25 13:38:15 crc kubenswrapper[4815]: I0225 13:38:15.808895 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-895cf5cf-5lnz6" podUID="3199ac65-2fd7-482f-b2fa-b5305ffa59d8" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: i/o timeout" Feb 25 13:38:16 crc kubenswrapper[4815]: I0225 13:38:16.360750 4815 generic.go:334] "Generic (PLEG): container finished" podID="5139ef47-6709-46f0-8f06-46cadbbb36b1" containerID="d09a900a7289126fcf20dcc95e26d94934df340a2b7e2ea2c89358cd26e4b697" exitCode=0 Feb 25 13:38:16 crc kubenswrapper[4815]: I0225 13:38:16.361308 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533778-xbfls" event={"ID":"5139ef47-6709-46f0-8f06-46cadbbb36b1","Type":"ContainerDied","Data":"d09a900a7289126fcf20dcc95e26d94934df340a2b7e2ea2c89358cd26e4b697"} Feb 25 13:38:16 crc kubenswrapper[4815]: I0225 13:38:16.366281 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"49be5892-2d7d-4e75-bf93-6a1d16e769fe","Type":"ContainerStarted","Data":"d27c3e23771676c3511779299bde6302a854d8941f0e8383263c10c3b0255c63"} Feb 25 13:38:16 crc kubenswrapper[4815]: I0225 13:38:16.371009 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"532b0818-f59f-4b9d-a4fd-6a976bcdb886","Type":"ContainerStarted","Data":"a2b8f88ce679426f3e4de1e8142e92559d8f4ccd4926b3b118b36d8249c419a7"} Feb 25 13:38:16 crc kubenswrapper[4815]: I0225 13:38:16.410722 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-645c577cc7-vpgnt"] Feb 25 13:38:16 crc kubenswrapper[4815]: W0225 13:38:16.410832 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod145d5de3_c56d_4209_b3b7_348f2f6d2365.slice/crio-ee07f1d14da2fcfdb7fcb74abe66b41f3176ad1f70fc65e0cc057f8220e04c03 WatchSource:0}: Error finding container ee07f1d14da2fcfdb7fcb74abe66b41f3176ad1f70fc65e0cc057f8220e04c03: Status 404 returned error can't find the container with id ee07f1d14da2fcfdb7fcb74abe66b41f3176ad1f70fc65e0cc057f8220e04c03 Feb 25 13:38:16 crc kubenswrapper[4815]: I0225 13:38:16.413149 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=13.413128349 podStartE2EDuration="13.413128349s" podCreationTimestamp="2026-02-25 13:38:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:16.41112182 +0000 UTC m=+1054.212219874" watchObservedRunningTime="2026-02-25 13:38:16.413128349 +0000 UTC m=+1054.214226403" Feb 25 13:38:17 crc kubenswrapper[4815]: I0225 13:38:17.385286 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3643568-6595-4c73-9c83-be6e28fabe83","Type":"ContainerStarted","Data":"df5eca668b03228f87a515ecd41160a6cf652cf136a652b575cfa3c9868694e9"} Feb 25 13:38:17 crc kubenswrapper[4815]: I0225 13:38:17.392274 4815 generic.go:334] "Generic (PLEG): container finished" podID="ee89a5e7-e5a4-412f-a52a-511dfca61265" containerID="c5a9af64e2d4e57b092b43e4a47fe7a5daffc4c0026054c3e700d2da1bb3823d" exitCode=0 Feb 25 13:38:17 crc kubenswrapper[4815]: I0225 13:38:17.392370 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mgztm" event={"ID":"ee89a5e7-e5a4-412f-a52a-511dfca61265","Type":"ContainerDied","Data":"c5a9af64e2d4e57b092b43e4a47fe7a5daffc4c0026054c3e700d2da1bb3823d"} Feb 25 13:38:17 crc kubenswrapper[4815]: I0225 13:38:17.396124 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"49be5892-2d7d-4e75-bf93-6a1d16e769fe","Type":"ContainerStarted","Data":"9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876"} Feb 25 13:38:17 crc kubenswrapper[4815]: I0225 13:38:17.396171 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"49be5892-2d7d-4e75-bf93-6a1d16e769fe","Type":"ContainerStarted","Data":"74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244"} Feb 25 13:38:17 crc kubenswrapper[4815]: I0225 13:38:17.400426 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-645c577cc7-vpgnt" event={"ID":"145d5de3-c56d-4209-b3b7-348f2f6d2365","Type":"ContainerStarted","Data":"97585b874331e8484826968cbdaea6a4929b28c335a1547f8af1086da98b9aad"} Feb 25 13:38:17 crc kubenswrapper[4815]: I0225 13:38:17.400475 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-645c577cc7-vpgnt" event={"ID":"145d5de3-c56d-4209-b3b7-348f2f6d2365","Type":"ContainerStarted","Data":"4c38128e88e041b8ab297bea5065cb3651ef99685d28363708fceec5f74ca125"} Feb 25 13:38:17 crc kubenswrapper[4815]: I0225 13:38:17.400488 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-645c577cc7-vpgnt" event={"ID":"145d5de3-c56d-4209-b3b7-348f2f6d2365","Type":"ContainerStarted","Data":"ee07f1d14da2fcfdb7fcb74abe66b41f3176ad1f70fc65e0cc057f8220e04c03"} Feb 25 13:38:17 crc kubenswrapper[4815]: I0225 13:38:17.442059 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-645c577cc7-vpgnt" podStartSLOduration=2.442037934 podStartE2EDuration="2.442037934s" podCreationTimestamp="2026-02-25 13:38:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:17.434729061 +0000 UTC m=+1055.235827115" watchObservedRunningTime="2026-02-25 13:38:17.442037934 +0000 UTC m=+1055.243135998" Feb 25 13:38:17 crc kubenswrapper[4815]: I0225 13:38:17.844057 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533778-xbfls" Feb 25 13:38:17 crc kubenswrapper[4815]: I0225 13:38:17.927800 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmzmb\" (UniqueName: \"kubernetes.io/projected/5139ef47-6709-46f0-8f06-46cadbbb36b1-kube-api-access-lmzmb\") pod \"5139ef47-6709-46f0-8f06-46cadbbb36b1\" (UID: \"5139ef47-6709-46f0-8f06-46cadbbb36b1\") " Feb 25 13:38:17 crc kubenswrapper[4815]: I0225 13:38:17.947245 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5139ef47-6709-46f0-8f06-46cadbbb36b1-kube-api-access-lmzmb" (OuterVolumeSpecName: "kube-api-access-lmzmb") pod "5139ef47-6709-46f0-8f06-46cadbbb36b1" (UID: "5139ef47-6709-46f0-8f06-46cadbbb36b1"). InnerVolumeSpecName "kube-api-access-lmzmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.029755 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmzmb\" (UniqueName: \"kubernetes.io/projected/5139ef47-6709-46f0-8f06-46cadbbb36b1-kube-api-access-lmzmb\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.426602 4815 generic.go:334] "Generic (PLEG): container finished" podID="63427264-5351-40c4-aeff-d35ea829ccd4" containerID="aa83b80cdebfae2c0db1d74c7475c986c41dd26e06d53c3e44db3c6c664fa6df" exitCode=0 Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.426876 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4x94p" event={"ID":"63427264-5351-40c4-aeff-d35ea829ccd4","Type":"ContainerDied","Data":"aa83b80cdebfae2c0db1d74c7475c986c41dd26e06d53c3e44db3c6c664fa6df"} Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.452489 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533778-xbfls" Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.453982 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533778-xbfls" event={"ID":"5139ef47-6709-46f0-8f06-46cadbbb36b1","Type":"ContainerDied","Data":"c74ec21c6d9c75a615e021af549d77630e6c23e5691db606f80ff9d7d94ac7d5"} Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.454030 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c74ec21c6d9c75a615e021af549d77630e6c23e5691db606f80ff9d7d94ac7d5" Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.454057 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.454361 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="49be5892-2d7d-4e75-bf93-6a1d16e769fe" containerName="glance-httpd" containerID="cri-o://9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876" gracePeriod=30 Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.454334 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="49be5892-2d7d-4e75-bf93-6a1d16e769fe" containerName="glance-log" containerID="cri-o://74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244" gracePeriod=30 Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.467847 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533772-htrc4"] Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.490926 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533772-htrc4"] Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.503271 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=29.503253924 podStartE2EDuration="29.503253924s" podCreationTimestamp="2026-02-25 13:37:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:18.495349183 +0000 UTC m=+1056.296447237" watchObservedRunningTime="2026-02-25 13:38:18.503253924 +0000 UTC m=+1056.304351978" Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.847405 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mgztm" Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.952208 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-combined-ca-bundle\") pod \"ee89a5e7-e5a4-412f-a52a-511dfca61265\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.952359 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-config-data\") pod \"ee89a5e7-e5a4-412f-a52a-511dfca61265\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.952418 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fzws\" (UniqueName: \"kubernetes.io/projected/ee89a5e7-e5a4-412f-a52a-511dfca61265-kube-api-access-7fzws\") pod \"ee89a5e7-e5a4-412f-a52a-511dfca61265\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.952492 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee89a5e7-e5a4-412f-a52a-511dfca61265-logs\") pod \"ee89a5e7-e5a4-412f-a52a-511dfca61265\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.952536 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-scripts\") pod \"ee89a5e7-e5a4-412f-a52a-511dfca61265\" (UID: \"ee89a5e7-e5a4-412f-a52a-511dfca61265\") " Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.953891 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee89a5e7-e5a4-412f-a52a-511dfca61265-logs" (OuterVolumeSpecName: "logs") pod "ee89a5e7-e5a4-412f-a52a-511dfca61265" (UID: "ee89a5e7-e5a4-412f-a52a-511dfca61265"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.959425 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee89a5e7-e5a4-412f-a52a-511dfca61265-kube-api-access-7fzws" (OuterVolumeSpecName: "kube-api-access-7fzws") pod "ee89a5e7-e5a4-412f-a52a-511dfca61265" (UID: "ee89a5e7-e5a4-412f-a52a-511dfca61265"). InnerVolumeSpecName "kube-api-access-7fzws". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.960160 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-scripts" (OuterVolumeSpecName: "scripts") pod "ee89a5e7-e5a4-412f-a52a-511dfca61265" (UID: "ee89a5e7-e5a4-412f-a52a-511dfca61265"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.962800 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a052dc18-c68b-4e41-8666-0d40b515d99b" path="/var/lib/kubelet/pods/a052dc18-c68b-4e41-8666-0d40b515d99b/volumes" Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.983731 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee89a5e7-e5a4-412f-a52a-511dfca61265" (UID: "ee89a5e7-e5a4-412f-a52a-511dfca61265"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:18 crc kubenswrapper[4815]: I0225 13:38:18.987695 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-config-data" (OuterVolumeSpecName: "config-data") pod "ee89a5e7-e5a4-412f-a52a-511dfca61265" (UID: "ee89a5e7-e5a4-412f-a52a-511dfca61265"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.054867 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee89a5e7-e5a4-412f-a52a-511dfca61265-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.054911 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.054942 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.054952 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee89a5e7-e5a4-412f-a52a-511dfca61265-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.054990 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fzws\" (UniqueName: \"kubernetes.io/projected/ee89a5e7-e5a4-412f-a52a-511dfca61265-kube-api-access-7fzws\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.172870 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.258787 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-scripts\") pod \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.259015 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-config-data\") pod \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.259079 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-public-tls-certs\") pod \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.259121 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df9xt\" (UniqueName: \"kubernetes.io/projected/49be5892-2d7d-4e75-bf93-6a1d16e769fe-kube-api-access-df9xt\") pod \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.259172 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49be5892-2d7d-4e75-bf93-6a1d16e769fe-logs\") pod \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.259204 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-combined-ca-bundle\") pod \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.259247 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.259271 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/49be5892-2d7d-4e75-bf93-6a1d16e769fe-httpd-run\") pod \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\" (UID: \"49be5892-2d7d-4e75-bf93-6a1d16e769fe\") " Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.260126 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49be5892-2d7d-4e75-bf93-6a1d16e769fe-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "49be5892-2d7d-4e75-bf93-6a1d16e769fe" (UID: "49be5892-2d7d-4e75-bf93-6a1d16e769fe"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.263143 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49be5892-2d7d-4e75-bf93-6a1d16e769fe-logs" (OuterVolumeSpecName: "logs") pod "49be5892-2d7d-4e75-bf93-6a1d16e769fe" (UID: "49be5892-2d7d-4e75-bf93-6a1d16e769fe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.264467 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49be5892-2d7d-4e75-bf93-6a1d16e769fe-kube-api-access-df9xt" (OuterVolumeSpecName: "kube-api-access-df9xt") pod "49be5892-2d7d-4e75-bf93-6a1d16e769fe" (UID: "49be5892-2d7d-4e75-bf93-6a1d16e769fe"). InnerVolumeSpecName "kube-api-access-df9xt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.264657 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-scripts" (OuterVolumeSpecName: "scripts") pod "49be5892-2d7d-4e75-bf93-6a1d16e769fe" (UID: "49be5892-2d7d-4e75-bf93-6a1d16e769fe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.269947 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "49be5892-2d7d-4e75-bf93-6a1d16e769fe" (UID: "49be5892-2d7d-4e75-bf93-6a1d16e769fe"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.290608 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "49be5892-2d7d-4e75-bf93-6a1d16e769fe" (UID: "49be5892-2d7d-4e75-bf93-6a1d16e769fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.311838 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-config-data" (OuterVolumeSpecName: "config-data") pod "49be5892-2d7d-4e75-bf93-6a1d16e769fe" (UID: "49be5892-2d7d-4e75-bf93-6a1d16e769fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.313798 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "49be5892-2d7d-4e75-bf93-6a1d16e769fe" (UID: "49be5892-2d7d-4e75-bf93-6a1d16e769fe"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.361581 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.361639 4815 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.361650 4815 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/49be5892-2d7d-4e75-bf93-6a1d16e769fe-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.361661 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.361916 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.361927 4815 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/49be5892-2d7d-4e75-bf93-6a1d16e769fe-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.361935 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df9xt\" (UniqueName: \"kubernetes.io/projected/49be5892-2d7d-4e75-bf93-6a1d16e769fe-kube-api-access-df9xt\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.361945 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49be5892-2d7d-4e75-bf93-6a1d16e769fe-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.378962 4815 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.463096 4815 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.466614 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mgztm" event={"ID":"ee89a5e7-e5a4-412f-a52a-511dfca61265","Type":"ContainerDied","Data":"c57f5520a69df9e29d9777d5cfc7db02cf80f6e84cdc3d18c36b829306d04eb8"} Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.466656 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c57f5520a69df9e29d9777d5cfc7db02cf80f6e84cdc3d18c36b829306d04eb8" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.466719 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mgztm" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.469282 4815 generic.go:334] "Generic (PLEG): container finished" podID="49be5892-2d7d-4e75-bf93-6a1d16e769fe" containerID="9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876" exitCode=0 Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.469310 4815 generic.go:334] "Generic (PLEG): container finished" podID="49be5892-2d7d-4e75-bf93-6a1d16e769fe" containerID="74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244" exitCode=143 Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.470172 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.478011 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"49be5892-2d7d-4e75-bf93-6a1d16e769fe","Type":"ContainerDied","Data":"9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876"} Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.478340 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"49be5892-2d7d-4e75-bf93-6a1d16e769fe","Type":"ContainerDied","Data":"74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244"} Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.478353 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"49be5892-2d7d-4e75-bf93-6a1d16e769fe","Type":"ContainerDied","Data":"d27c3e23771676c3511779299bde6302a854d8941f0e8383263c10c3b0255c63"} Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.478368 4815 scope.go:117] "RemoveContainer" containerID="9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.506569 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.517352 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.520838 4815 scope.go:117] "RemoveContainer" containerID="74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.541632 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:38:19 crc kubenswrapper[4815]: E0225 13:38:19.541977 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee89a5e7-e5a4-412f-a52a-511dfca61265" containerName="placement-db-sync" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.541991 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee89a5e7-e5a4-412f-a52a-511dfca61265" containerName="placement-db-sync" Feb 25 13:38:19 crc kubenswrapper[4815]: E0225 13:38:19.542005 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5139ef47-6709-46f0-8f06-46cadbbb36b1" containerName="oc" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.542012 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="5139ef47-6709-46f0-8f06-46cadbbb36b1" containerName="oc" Feb 25 13:38:19 crc kubenswrapper[4815]: E0225 13:38:19.542027 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49be5892-2d7d-4e75-bf93-6a1d16e769fe" containerName="glance-httpd" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.542033 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="49be5892-2d7d-4e75-bf93-6a1d16e769fe" containerName="glance-httpd" Feb 25 13:38:19 crc kubenswrapper[4815]: E0225 13:38:19.542042 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49be5892-2d7d-4e75-bf93-6a1d16e769fe" containerName="glance-log" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.542048 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="49be5892-2d7d-4e75-bf93-6a1d16e769fe" containerName="glance-log" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.542217 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee89a5e7-e5a4-412f-a52a-511dfca61265" containerName="placement-db-sync" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.542247 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="49be5892-2d7d-4e75-bf93-6a1d16e769fe" containerName="glance-log" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.542269 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="5139ef47-6709-46f0-8f06-46cadbbb36b1" containerName="oc" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.542307 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="49be5892-2d7d-4e75-bf93-6a1d16e769fe" containerName="glance-httpd" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.543131 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.545488 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.545639 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.553671 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.556616 4815 scope.go:117] "RemoveContainer" containerID="9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876" Feb 25 13:38:19 crc kubenswrapper[4815]: E0225 13:38:19.559693 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876\": container with ID starting with 9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876 not found: ID does not exist" containerID="9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.559725 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876"} err="failed to get container status \"9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876\": rpc error: code = NotFound desc = could not find container \"9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876\": container with ID starting with 9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876 not found: ID does not exist" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.559746 4815 scope.go:117] "RemoveContainer" containerID="74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244" Feb 25 13:38:19 crc kubenswrapper[4815]: E0225 13:38:19.560113 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244\": container with ID starting with 74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244 not found: ID does not exist" containerID="74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.560133 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244"} err="failed to get container status \"74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244\": rpc error: code = NotFound desc = could not find container \"74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244\": container with ID starting with 74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244 not found: ID does not exist" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.560146 4815 scope.go:117] "RemoveContainer" containerID="9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.560458 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876"} err="failed to get container status \"9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876\": rpc error: code = NotFound desc = could not find container \"9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876\": container with ID starting with 9d96773bd56385b178a7644e7190034e9b4b91347a0c85f3c1fe7db0e8ed5876 not found: ID does not exist" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.560475 4815 scope.go:117] "RemoveContainer" containerID="74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.562067 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244"} err="failed to get container status \"74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244\": rpc error: code = NotFound desc = could not find container \"74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244\": container with ID starting with 74927e82cd2d98cc4e085ff356560ae91c258cf95e5bc9ac2a4a3d4e9423f244 not found: ID does not exist" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.631564 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-75dc9744b8-t8qpj"] Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.632988 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.635939 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.640282 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-2z6dv" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.640551 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.640666 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.648707 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-75dc9744b8-t8qpj"] Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.648884 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.666533 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.666615 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-config-data\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.666637 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpn4z\" (UniqueName: \"kubernetes.io/projected/96b0bb62-f8d6-4253-a70a-a95f071472df-kube-api-access-dpn4z\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.666661 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96b0bb62-f8d6-4253-a70a-a95f071472df-logs\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.666691 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/96b0bb62-f8d6-4253-a70a-a95f071472df-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.666708 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.666729 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-scripts\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.666753 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.768239 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.768330 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nv4p\" (UniqueName: \"kubernetes.io/projected/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-kube-api-access-6nv4p\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.768354 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-config-data\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.768372 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpn4z\" (UniqueName: \"kubernetes.io/projected/96b0bb62-f8d6-4253-a70a-a95f071472df-kube-api-access-dpn4z\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.768395 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96b0bb62-f8d6-4253-a70a-a95f071472df-logs\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.768424 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-logs\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.768443 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/96b0bb62-f8d6-4253-a70a-a95f071472df-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.768458 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.768473 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-scripts\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.768489 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-public-tls-certs\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.768520 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-internal-tls-certs\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.768539 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.768647 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-scripts\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.768699 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-config-data\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.768749 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-combined-ca-bundle\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.769336 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96b0bb62-f8d6-4253-a70a-a95f071472df-logs\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.769560 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/96b0bb62-f8d6-4253-a70a-a95f071472df-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.769824 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.773754 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.779266 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-scripts\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.781590 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-config-data\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.784275 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.807159 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpn4z\" (UniqueName: \"kubernetes.io/projected/96b0bb62-f8d6-4253-a70a-a95f071472df-kube-api-access-dpn4z\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.824943 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.863549 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.870266 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-config-data\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.870294 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-scripts\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.870319 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-combined-ca-bundle\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.870384 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nv4p\" (UniqueName: \"kubernetes.io/projected/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-kube-api-access-6nv4p\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.870446 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-logs\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.870468 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-public-tls-certs\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.870488 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-internal-tls-certs\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.877187 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-internal-tls-certs\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.877737 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-logs\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.884048 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-config-data\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.886202 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-public-tls-certs\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.886202 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-combined-ca-bundle\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.886473 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-scripts\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.911145 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nv4p\" (UniqueName: \"kubernetes.io/projected/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-kube-api-access-6nv4p\") pod \"placement-75dc9744b8-t8qpj\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:19 crc kubenswrapper[4815]: I0225 13:38:19.955945 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:20 crc kubenswrapper[4815]: I0225 13:38:20.949009 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49be5892-2d7d-4e75-bf93-6a1d16e769fe" path="/var/lib/kubelet/pods/49be5892-2d7d-4e75-bf93-6a1d16e769fe/volumes" Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.518015 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4x94p" event={"ID":"63427264-5351-40c4-aeff-d35ea829ccd4","Type":"ContainerDied","Data":"65120763d9188de1d47032dd07435216a8381d897169e3f1e152f2bf2ce920e7"} Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.518058 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65120763d9188de1d47032dd07435216a8381d897169e3f1e152f2bf2ce920e7" Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.542101 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.606067 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-credential-keys\") pod \"63427264-5351-40c4-aeff-d35ea829ccd4\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.606119 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-scripts\") pod \"63427264-5351-40c4-aeff-d35ea829ccd4\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.606171 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrpwb\" (UniqueName: \"kubernetes.io/projected/63427264-5351-40c4-aeff-d35ea829ccd4-kube-api-access-wrpwb\") pod \"63427264-5351-40c4-aeff-d35ea829ccd4\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.606281 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-combined-ca-bundle\") pod \"63427264-5351-40c4-aeff-d35ea829ccd4\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.606325 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-fernet-keys\") pod \"63427264-5351-40c4-aeff-d35ea829ccd4\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.606366 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-config-data\") pod \"63427264-5351-40c4-aeff-d35ea829ccd4\" (UID: \"63427264-5351-40c4-aeff-d35ea829ccd4\") " Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.612102 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-scripts" (OuterVolumeSpecName: "scripts") pod "63427264-5351-40c4-aeff-d35ea829ccd4" (UID: "63427264-5351-40c4-aeff-d35ea829ccd4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.613694 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "63427264-5351-40c4-aeff-d35ea829ccd4" (UID: "63427264-5351-40c4-aeff-d35ea829ccd4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.617667 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63427264-5351-40c4-aeff-d35ea829ccd4-kube-api-access-wrpwb" (OuterVolumeSpecName: "kube-api-access-wrpwb") pod "63427264-5351-40c4-aeff-d35ea829ccd4" (UID: "63427264-5351-40c4-aeff-d35ea829ccd4"). InnerVolumeSpecName "kube-api-access-wrpwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.627582 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "63427264-5351-40c4-aeff-d35ea829ccd4" (UID: "63427264-5351-40c4-aeff-d35ea829ccd4"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.643601 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63427264-5351-40c4-aeff-d35ea829ccd4" (UID: "63427264-5351-40c4-aeff-d35ea829ccd4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.661092 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-config-data" (OuterVolumeSpecName: "config-data") pod "63427264-5351-40c4-aeff-d35ea829ccd4" (UID: "63427264-5351-40c4-aeff-d35ea829ccd4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.708919 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.708963 4815 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.708975 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.708986 4815 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.708996 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63427264-5351-40c4-aeff-d35ea829ccd4-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:21 crc kubenswrapper[4815]: I0225 13:38:21.709007 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrpwb\" (UniqueName: \"kubernetes.io/projected/63427264-5351-40c4-aeff-d35ea829ccd4-kube-api-access-wrpwb\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.526077 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4x94p" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.730098 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6cdd69d876-dmcfv"] Feb 25 13:38:22 crc kubenswrapper[4815]: E0225 13:38:22.730431 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63427264-5351-40c4-aeff-d35ea829ccd4" containerName="keystone-bootstrap" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.730447 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="63427264-5351-40c4-aeff-d35ea829ccd4" containerName="keystone-bootstrap" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.730635 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="63427264-5351-40c4-aeff-d35ea829ccd4" containerName="keystone-bootstrap" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.731165 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.736296 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.736492 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.736649 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.736808 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.736822 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-tftcl" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.736960 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.745335 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6cdd69d876-dmcfv"] Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.832969 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-scripts\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.833012 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-credential-keys\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.833031 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-config-data\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.833180 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9zbb\" (UniqueName: \"kubernetes.io/projected/9fae323c-a7e9-46a6-b658-1c81c99240e1-kube-api-access-v9zbb\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.833275 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-fernet-keys\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.833379 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-public-tls-certs\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.833480 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-internal-tls-certs\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.833622 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-combined-ca-bundle\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.936647 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-scripts\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.936708 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-credential-keys\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.936735 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-config-data\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.936775 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9zbb\" (UniqueName: \"kubernetes.io/projected/9fae323c-a7e9-46a6-b658-1c81c99240e1-kube-api-access-v9zbb\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.936812 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-fernet-keys\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.936881 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-public-tls-certs\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.936938 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-internal-tls-certs\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.936967 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-combined-ca-bundle\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.941580 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-scripts\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.941671 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-combined-ca-bundle\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.942250 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-fernet-keys\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.943951 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-credential-keys\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.945019 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-internal-tls-certs\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.945334 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-public-tls-certs\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.945375 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fae323c-a7e9-46a6-b658-1c81c99240e1-config-data\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:22 crc kubenswrapper[4815]: I0225 13:38:22.958885 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9zbb\" (UniqueName: \"kubernetes.io/projected/9fae323c-a7e9-46a6-b658-1c81c99240e1-kube-api-access-v9zbb\") pod \"keystone-6cdd69d876-dmcfv\" (UID: \"9fae323c-a7e9-46a6-b658-1c81c99240e1\") " pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.052955 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.208683 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.346105 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-lbk4j"] Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.346668 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" podUID="73a1f5fc-5524-4753-9c6d-87c8d21f592a" containerName="dnsmasq-dns" containerID="cri-o://5ad820c001b1adaa66f858b1dff7082c42d1a32530cbe5680e51e85725e39f48" gracePeriod=10 Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.370144 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.501656 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.501702 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.561548 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.561616 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.572098 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3643568-6595-4c73-9c83-be6e28fabe83","Type":"ContainerStarted","Data":"431e99a3dca863f8134146b4183f20a39dd63d734e655e385c0aceb2f42d819f"} Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.574098 4815 generic.go:334] "Generic (PLEG): container finished" podID="73a1f5fc-5524-4753-9c6d-87c8d21f592a" containerID="5ad820c001b1adaa66f858b1dff7082c42d1a32530cbe5680e51e85725e39f48" exitCode=0 Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.575244 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" event={"ID":"73a1f5fc-5524-4753-9c6d-87c8d21f592a","Type":"ContainerDied","Data":"5ad820c001b1adaa66f858b1dff7082c42d1a32530cbe5680e51e85725e39f48"} Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.575274 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.575370 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.691848 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.766946 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-75dc9744b8-t8qpj"] Feb 25 13:38:23 crc kubenswrapper[4815]: W0225 13:38:23.775438 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafdec2fd_f5ed_49ce_8ce0_4e5b0e2b7fe4.slice/crio-ee6a4dcb6b577fb8a0d21147dce77741bb4cad2771cab8604d803f2efc4271f9 WatchSource:0}: Error finding container ee6a4dcb6b577fb8a0d21147dce77741bb4cad2771cab8604d803f2efc4271f9: Status 404 returned error can't find the container with id ee6a4dcb6b577fb8a0d21147dce77741bb4cad2771cab8604d803f2efc4271f9 Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.838034 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.839407 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.921353 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6cdd69d876-dmcfv"] Feb 25 13:38:23 crc kubenswrapper[4815]: W0225 13:38:23.926603 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fae323c_a7e9_46a6_b658_1c81c99240e1.slice/crio-754ad5442f9904a2b64983f6b2efd73f0dbf496fa128e62bee47ebcaa36f508b WatchSource:0}: Error finding container 754ad5442f9904a2b64983f6b2efd73f0dbf496fa128e62bee47ebcaa36f508b: Status 404 returned error can't find the container with id 754ad5442f9904a2b64983f6b2efd73f0dbf496fa128e62bee47ebcaa36f508b Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.979731 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.980109 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:38:23 crc kubenswrapper[4815]: I0225 13:38:23.981572 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7657655688-k9k6t" podUID="f10b8779-9d5e-427d-847c-3250816e89b9" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.061738 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.195274 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-ovsdbserver-sb\") pod \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.195356 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-ovsdbserver-nb\") pod \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.195463 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-dns-swift-storage-0\") pod \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.195545 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-dns-svc\") pod \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.195662 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-config\") pod \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.195758 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz6rw\" (UniqueName: \"kubernetes.io/projected/73a1f5fc-5524-4753-9c6d-87c8d21f592a-kube-api-access-nz6rw\") pod \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\" (UID: \"73a1f5fc-5524-4753-9c6d-87c8d21f592a\") " Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.217797 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73a1f5fc-5524-4753-9c6d-87c8d21f592a-kube-api-access-nz6rw" (OuterVolumeSpecName: "kube-api-access-nz6rw") pod "73a1f5fc-5524-4753-9c6d-87c8d21f592a" (UID: "73a1f5fc-5524-4753-9c6d-87c8d21f592a"). InnerVolumeSpecName "kube-api-access-nz6rw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.298624 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz6rw\" (UniqueName: \"kubernetes.io/projected/73a1f5fc-5524-4753-9c6d-87c8d21f592a-kube-api-access-nz6rw\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.473783 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "73a1f5fc-5524-4753-9c6d-87c8d21f592a" (UID: "73a1f5fc-5524-4753-9c6d-87c8d21f592a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.480322 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-config" (OuterVolumeSpecName: "config") pod "73a1f5fc-5524-4753-9c6d-87c8d21f592a" (UID: "73a1f5fc-5524-4753-9c6d-87c8d21f592a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.492143 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "73a1f5fc-5524-4753-9c6d-87c8d21f592a" (UID: "73a1f5fc-5524-4753-9c6d-87c8d21f592a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.497590 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "73a1f5fc-5524-4753-9c6d-87c8d21f592a" (UID: "73a1f5fc-5524-4753-9c6d-87c8d21f592a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.504299 4815 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.504343 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.504387 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.504400 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.513420 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "73a1f5fc-5524-4753-9c6d-87c8d21f592a" (UID: "73a1f5fc-5524-4753-9c6d-87c8d21f592a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.587245 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" event={"ID":"73a1f5fc-5524-4753-9c6d-87c8d21f592a","Type":"ContainerDied","Data":"6f3785a779335d37182644a434564d3b739abdacced322230a12f2aec88fd60c"} Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.587300 4815 scope.go:117] "RemoveContainer" containerID="5ad820c001b1adaa66f858b1dff7082c42d1a32530cbe5680e51e85725e39f48" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.587438 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-lbk4j" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.590923 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6cdd69d876-dmcfv" event={"ID":"9fae323c-a7e9-46a6-b658-1c81c99240e1","Type":"ContainerStarted","Data":"637e85cfefe965eb89e615cb4e4bd548a8d2ae14d1d7e4ffd97d35ea46d5ad5b"} Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.590967 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6cdd69d876-dmcfv" event={"ID":"9fae323c-a7e9-46a6-b658-1c81c99240e1","Type":"ContainerStarted","Data":"754ad5442f9904a2b64983f6b2efd73f0dbf496fa128e62bee47ebcaa36f508b"} Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.591901 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.594460 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"96b0bb62-f8d6-4253-a70a-a95f071472df","Type":"ContainerStarted","Data":"651b85a5d30918624c3825b1ff0fb632136bebcff4c904e1fdef8baed1491991"} Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.602227 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75dc9744b8-t8qpj" event={"ID":"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4","Type":"ContainerStarted","Data":"68c5a50e0da94b36c79a7d43ce763c4da7f48c6c14c0fcd8ec3edbb5a3a77956"} Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.602278 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75dc9744b8-t8qpj" event={"ID":"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4","Type":"ContainerStarted","Data":"ee6a4dcb6b577fb8a0d21147dce77741bb4cad2771cab8604d803f2efc4271f9"} Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.608746 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73a1f5fc-5524-4753-9c6d-87c8d21f592a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.633958 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6cdd69d876-dmcfv" podStartSLOduration=2.633941865 podStartE2EDuration="2.633941865s" podCreationTimestamp="2026-02-25 13:38:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:24.61973759 +0000 UTC m=+1062.420835644" watchObservedRunningTime="2026-02-25 13:38:24.633941865 +0000 UTC m=+1062.435039919" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.643750 4815 scope.go:117] "RemoveContainer" containerID="72af52454ed4764867fb2a872f2680abe7428323592a8731116c568dbf259e4d" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.648989 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-lbk4j"] Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.655446 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-lbk4j"] Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.771029 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-b695645c6-vszf4"] Feb 25 13:38:24 crc kubenswrapper[4815]: E0225 13:38:24.771363 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73a1f5fc-5524-4753-9c6d-87c8d21f592a" containerName="init" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.771374 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="73a1f5fc-5524-4753-9c6d-87c8d21f592a" containerName="init" Feb 25 13:38:24 crc kubenswrapper[4815]: E0225 13:38:24.771385 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73a1f5fc-5524-4753-9c6d-87c8d21f592a" containerName="dnsmasq-dns" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.771392 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="73a1f5fc-5524-4753-9c6d-87c8d21f592a" containerName="dnsmasq-dns" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.771582 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="73a1f5fc-5524-4753-9c6d-87c8d21f592a" containerName="dnsmasq-dns" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.772410 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.789750 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b695645c6-vszf4"] Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.814549 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99ea30bb-39c5-47ee-aa2a-008fe013de18-public-tls-certs\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.814863 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99ea30bb-39c5-47ee-aa2a-008fe013de18-scripts\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.814983 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ea30bb-39c5-47ee-aa2a-008fe013de18-combined-ca-bundle\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.815096 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99ea30bb-39c5-47ee-aa2a-008fe013de18-logs\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.815209 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99ea30bb-39c5-47ee-aa2a-008fe013de18-internal-tls-certs\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.815340 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrcwt\" (UniqueName: \"kubernetes.io/projected/99ea30bb-39c5-47ee-aa2a-008fe013de18-kube-api-access-xrcwt\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.815466 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ea30bb-39c5-47ee-aa2a-008fe013de18-config-data\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.917169 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99ea30bb-39c5-47ee-aa2a-008fe013de18-public-tls-certs\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.917227 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99ea30bb-39c5-47ee-aa2a-008fe013de18-scripts\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.917257 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ea30bb-39c5-47ee-aa2a-008fe013de18-combined-ca-bundle\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.917270 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99ea30bb-39c5-47ee-aa2a-008fe013de18-logs\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.917303 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99ea30bb-39c5-47ee-aa2a-008fe013de18-internal-tls-certs\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.917331 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrcwt\" (UniqueName: \"kubernetes.io/projected/99ea30bb-39c5-47ee-aa2a-008fe013de18-kube-api-access-xrcwt\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.917353 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ea30bb-39c5-47ee-aa2a-008fe013de18-config-data\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.918922 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99ea30bb-39c5-47ee-aa2a-008fe013de18-logs\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.921265 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ea30bb-39c5-47ee-aa2a-008fe013de18-config-data\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.924026 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99ea30bb-39c5-47ee-aa2a-008fe013de18-scripts\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.924239 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99ea30bb-39c5-47ee-aa2a-008fe013de18-public-tls-certs\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.927078 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ea30bb-39c5-47ee-aa2a-008fe013de18-combined-ca-bundle\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.930132 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/99ea30bb-39c5-47ee-aa2a-008fe013de18-internal-tls-certs\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.945928 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73a1f5fc-5524-4753-9c6d-87c8d21f592a" path="/var/lib/kubelet/pods/73a1f5fc-5524-4753-9c6d-87c8d21f592a/volumes" Feb 25 13:38:24 crc kubenswrapper[4815]: I0225 13:38:24.964130 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrcwt\" (UniqueName: \"kubernetes.io/projected/99ea30bb-39c5-47ee-aa2a-008fe013de18-kube-api-access-xrcwt\") pod \"placement-b695645c6-vszf4\" (UID: \"99ea30bb-39c5-47ee-aa2a-008fe013de18\") " pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:25 crc kubenswrapper[4815]: I0225 13:38:25.106017 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:25 crc kubenswrapper[4815]: I0225 13:38:25.611560 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75dc9744b8-t8qpj" event={"ID":"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4","Type":"ContainerStarted","Data":"2aac1dd5a8902bfcc26014020926a1705bb21bc2b6daa68d882f15c2e47a07a2"} Feb 25 13:38:25 crc kubenswrapper[4815]: I0225 13:38:25.612018 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:25 crc kubenswrapper[4815]: I0225 13:38:25.612043 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:25 crc kubenswrapper[4815]: I0225 13:38:25.615665 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"96b0bb62-f8d6-4253-a70a-a95f071472df","Type":"ContainerStarted","Data":"bf90c6f53416fb5d0819b337fee413c8a1085ef1909bae7fb50aa793a4a122af"} Feb 25 13:38:25 crc kubenswrapper[4815]: I0225 13:38:25.615711 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"96b0bb62-f8d6-4253-a70a-a95f071472df","Type":"ContainerStarted","Data":"9fbfb6af1d9b2620d8d826b615a233bc26850af324ad4bc7b503385bcffffea4"} Feb 25 13:38:25 crc kubenswrapper[4815]: I0225 13:38:25.615742 4815 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 25 13:38:25 crc kubenswrapper[4815]: I0225 13:38:25.615756 4815 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 25 13:38:25 crc kubenswrapper[4815]: I0225 13:38:25.637781 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-75dc9744b8-t8qpj" podStartSLOduration=6.637762856 podStartE2EDuration="6.637762856s" podCreationTimestamp="2026-02-25 13:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:25.629398662 +0000 UTC m=+1063.430496716" watchObservedRunningTime="2026-02-25 13:38:25.637762856 +0000 UTC m=+1063.438860910" Feb 25 13:38:25 crc kubenswrapper[4815]: I0225 13:38:25.659735 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.659720399 podStartE2EDuration="6.659720399s" podCreationTimestamp="2026-02-25 13:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:25.6532606 +0000 UTC m=+1063.454358654" watchObservedRunningTime="2026-02-25 13:38:25.659720399 +0000 UTC m=+1063.460818453" Feb 25 13:38:25 crc kubenswrapper[4815]: W0225 13:38:25.685438 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99ea30bb_39c5_47ee_aa2a_008fe013de18.slice/crio-8d27a7ede0fad81f93cfd04737ebddfe060e8ccdffd2c0a4defc0b7a89120b5c WatchSource:0}: Error finding container 8d27a7ede0fad81f93cfd04737ebddfe060e8ccdffd2c0a4defc0b7a89120b5c: Status 404 returned error can't find the container with id 8d27a7ede0fad81f93cfd04737ebddfe060e8ccdffd2c0a4defc0b7a89120b5c Feb 25 13:38:25 crc kubenswrapper[4815]: I0225 13:38:25.693884 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b695645c6-vszf4"] Feb 25 13:38:26 crc kubenswrapper[4815]: I0225 13:38:26.147211 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 25 13:38:26 crc kubenswrapper[4815]: I0225 13:38:26.147869 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 25 13:38:26 crc kubenswrapper[4815]: I0225 13:38:26.625588 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b695645c6-vszf4" event={"ID":"99ea30bb-39c5-47ee-aa2a-008fe013de18","Type":"ContainerStarted","Data":"8d27a7ede0fad81f93cfd04737ebddfe060e8ccdffd2c0a4defc0b7a89120b5c"} Feb 25 13:38:27 crc kubenswrapper[4815]: I0225 13:38:27.641084 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b695645c6-vszf4" event={"ID":"99ea30bb-39c5-47ee-aa2a-008fe013de18","Type":"ContainerStarted","Data":"095f8ae05a415fab3cde1a1577114713fb6e0ac02c87a06852d1cddd55459063"} Feb 25 13:38:27 crc kubenswrapper[4815]: I0225 13:38:27.641497 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b695645c6-vszf4" event={"ID":"99ea30bb-39c5-47ee-aa2a-008fe013de18","Type":"ContainerStarted","Data":"aa545b3aa868f1dfd8b50963e8f506f1dc54e9fdc1d643a285531c6880790bce"} Feb 25 13:38:27 crc kubenswrapper[4815]: I0225 13:38:27.641553 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:27 crc kubenswrapper[4815]: I0225 13:38:27.641574 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:27 crc kubenswrapper[4815]: I0225 13:38:27.680037 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-b695645c6-vszf4" podStartSLOduration=3.680011702 podStartE2EDuration="3.680011702s" podCreationTimestamp="2026-02-25 13:38:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:27.665806387 +0000 UTC m=+1065.466904461" watchObservedRunningTime="2026-02-25 13:38:27.680011702 +0000 UTC m=+1065.481109796" Feb 25 13:38:29 crc kubenswrapper[4815]: I0225 13:38:29.865051 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 25 13:38:29 crc kubenswrapper[4815]: I0225 13:38:29.865442 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 25 13:38:29 crc kubenswrapper[4815]: I0225 13:38:29.914410 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 25 13:38:29 crc kubenswrapper[4815]: I0225 13:38:29.916743 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 25 13:38:30 crc kubenswrapper[4815]: I0225 13:38:30.667884 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 25 13:38:30 crc kubenswrapper[4815]: I0225 13:38:30.668146 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 25 13:38:32 crc kubenswrapper[4815]: I0225 13:38:32.598380 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 25 13:38:32 crc kubenswrapper[4815]: I0225 13:38:32.626045 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 25 13:38:32 crc kubenswrapper[4815]: I0225 13:38:32.691797 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-d4ssg" event={"ID":"7bdc9071-e863-4108-9234-4f201732f435","Type":"ContainerStarted","Data":"9d5688a7d91b39b94743229dfb65ac8852aa764ec59864066e8ded936e9d1ece"} Feb 25 13:38:32 crc kubenswrapper[4815]: I0225 13:38:32.707045 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f3643568-6595-4c73-9c83-be6e28fabe83" containerName="ceilometer-central-agent" containerID="cri-o://dc7434d19b0960010335e00c9621ca0164b153f5cd542c93cb16504e4883f085" gracePeriod=30 Feb 25 13:38:32 crc kubenswrapper[4815]: I0225 13:38:32.707135 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3643568-6595-4c73-9c83-be6e28fabe83","Type":"ContainerStarted","Data":"47cb76120db6d64fb0e087e27d5cbbb53191df2db146504b0d521e66c55ec08f"} Feb 25 13:38:32 crc kubenswrapper[4815]: I0225 13:38:32.707168 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 13:38:32 crc kubenswrapper[4815]: I0225 13:38:32.707470 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f3643568-6595-4c73-9c83-be6e28fabe83" containerName="proxy-httpd" containerID="cri-o://47cb76120db6d64fb0e087e27d5cbbb53191df2db146504b0d521e66c55ec08f" gracePeriod=30 Feb 25 13:38:32 crc kubenswrapper[4815]: I0225 13:38:32.707533 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f3643568-6595-4c73-9c83-be6e28fabe83" containerName="sg-core" containerID="cri-o://431e99a3dca863f8134146b4183f20a39dd63d734e655e385c0aceb2f42d819f" gracePeriod=30 Feb 25 13:38:32 crc kubenswrapper[4815]: I0225 13:38:32.707568 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f3643568-6595-4c73-9c83-be6e28fabe83" containerName="ceilometer-notification-agent" containerID="cri-o://df5eca668b03228f87a515ecd41160a6cf652cf136a652b575cfa3c9868694e9" gracePeriod=30 Feb 25 13:38:32 crc kubenswrapper[4815]: I0225 13:38:32.719645 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-d4ssg" podStartSLOduration=2.600988432 podStartE2EDuration="48.719630562s" podCreationTimestamp="2026-02-25 13:37:44 +0000 UTC" firstStartedPulling="2026-02-25 13:37:46.073470092 +0000 UTC m=+1023.874568146" lastFinishedPulling="2026-02-25 13:38:32.192112222 +0000 UTC m=+1069.993210276" observedRunningTime="2026-02-25 13:38:32.70349022 +0000 UTC m=+1070.504588274" watchObservedRunningTime="2026-02-25 13:38:32.719630562 +0000 UTC m=+1070.520728616" Feb 25 13:38:32 crc kubenswrapper[4815]: I0225 13:38:32.737887 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.812006163 podStartE2EDuration="48.737870935s" podCreationTimestamp="2026-02-25 13:37:44 +0000 UTC" firstStartedPulling="2026-02-25 13:37:45.401669789 +0000 UTC m=+1023.202767843" lastFinishedPulling="2026-02-25 13:38:32.327534541 +0000 UTC m=+1070.128632615" observedRunningTime="2026-02-25 13:38:32.728315316 +0000 UTC m=+1070.529413370" watchObservedRunningTime="2026-02-25 13:38:32.737870935 +0000 UTC m=+1070.538968989" Feb 25 13:38:33 crc kubenswrapper[4815]: I0225 13:38:33.716337 4815 generic.go:334] "Generic (PLEG): container finished" podID="f3643568-6595-4c73-9c83-be6e28fabe83" containerID="47cb76120db6d64fb0e087e27d5cbbb53191df2db146504b0d521e66c55ec08f" exitCode=0 Feb 25 13:38:33 crc kubenswrapper[4815]: I0225 13:38:33.716594 4815 generic.go:334] "Generic (PLEG): container finished" podID="f3643568-6595-4c73-9c83-be6e28fabe83" containerID="431e99a3dca863f8134146b4183f20a39dd63d734e655e385c0aceb2f42d819f" exitCode=2 Feb 25 13:38:33 crc kubenswrapper[4815]: I0225 13:38:33.716395 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3643568-6595-4c73-9c83-be6e28fabe83","Type":"ContainerDied","Data":"47cb76120db6d64fb0e087e27d5cbbb53191df2db146504b0d521e66c55ec08f"} Feb 25 13:38:33 crc kubenswrapper[4815]: I0225 13:38:33.716642 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3643568-6595-4c73-9c83-be6e28fabe83","Type":"ContainerDied","Data":"431e99a3dca863f8134146b4183f20a39dd63d734e655e385c0aceb2f42d819f"} Feb 25 13:38:33 crc kubenswrapper[4815]: I0225 13:38:33.716656 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3643568-6595-4c73-9c83-be6e28fabe83","Type":"ContainerDied","Data":"dc7434d19b0960010335e00c9621ca0164b153f5cd542c93cb16504e4883f085"} Feb 25 13:38:33 crc kubenswrapper[4815]: I0225 13:38:33.716605 4815 generic.go:334] "Generic (PLEG): container finished" podID="f3643568-6595-4c73-9c83-be6e28fabe83" containerID="dc7434d19b0960010335e00c9621ca0164b153f5cd542c93cb16504e4883f085" exitCode=0 Feb 25 13:38:33 crc kubenswrapper[4815]: I0225 13:38:33.718585 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-576gs" event={"ID":"04e3d721-2179-4457-8cc6-72af3b3f9257","Type":"ContainerStarted","Data":"4ce2c3c9b64d5f047c35dc1e7ac70a87bf7be1bc271d659eef2e7ae14c2e1a8e"} Feb 25 13:38:33 crc kubenswrapper[4815]: I0225 13:38:33.737589 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-576gs" podStartSLOduration=3.193718161 podStartE2EDuration="49.737573586s" podCreationTimestamp="2026-02-25 13:37:44 +0000 UTC" firstStartedPulling="2026-02-25 13:37:45.736471306 +0000 UTC m=+1023.537569360" lastFinishedPulling="2026-02-25 13:38:32.280326701 +0000 UTC m=+1070.081424785" observedRunningTime="2026-02-25 13:38:33.730942083 +0000 UTC m=+1071.532040137" watchObservedRunningTime="2026-02-25 13:38:33.737573586 +0000 UTC m=+1071.538671640" Feb 25 13:38:33 crc kubenswrapper[4815]: I0225 13:38:33.838628 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6497d456d-s5g59" podUID="5d076fbb-4499-456f-82ee-211a8909fc99" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 25 13:38:33 crc kubenswrapper[4815]: I0225 13:38:33.979316 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7657655688-k9k6t" podUID="f10b8779-9d5e-427d-847c-3250816e89b9" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.498287 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.605397 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-sg-core-conf-yaml\") pod \"f3643568-6595-4c73-9c83-be6e28fabe83\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.605708 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-config-data\") pod \"f3643568-6595-4c73-9c83-be6e28fabe83\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.605734 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-combined-ca-bundle\") pod \"f3643568-6595-4c73-9c83-be6e28fabe83\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.605782 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mv9lz\" (UniqueName: \"kubernetes.io/projected/f3643568-6595-4c73-9c83-be6e28fabe83-kube-api-access-mv9lz\") pod \"f3643568-6595-4c73-9c83-be6e28fabe83\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.605864 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-scripts\") pod \"f3643568-6595-4c73-9c83-be6e28fabe83\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.605890 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3643568-6595-4c73-9c83-be6e28fabe83-run-httpd\") pod \"f3643568-6595-4c73-9c83-be6e28fabe83\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.606008 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3643568-6595-4c73-9c83-be6e28fabe83-log-httpd\") pod \"f3643568-6595-4c73-9c83-be6e28fabe83\" (UID: \"f3643568-6595-4c73-9c83-be6e28fabe83\") " Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.606677 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3643568-6595-4c73-9c83-be6e28fabe83-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f3643568-6595-4c73-9c83-be6e28fabe83" (UID: "f3643568-6595-4c73-9c83-be6e28fabe83"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.608028 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3643568-6595-4c73-9c83-be6e28fabe83-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f3643568-6595-4c73-9c83-be6e28fabe83" (UID: "f3643568-6595-4c73-9c83-be6e28fabe83"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.614623 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-scripts" (OuterVolumeSpecName: "scripts") pod "f3643568-6595-4c73-9c83-be6e28fabe83" (UID: "f3643568-6595-4c73-9c83-be6e28fabe83"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.626799 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3643568-6595-4c73-9c83-be6e28fabe83-kube-api-access-mv9lz" (OuterVolumeSpecName: "kube-api-access-mv9lz") pod "f3643568-6595-4c73-9c83-be6e28fabe83" (UID: "f3643568-6595-4c73-9c83-be6e28fabe83"). InnerVolumeSpecName "kube-api-access-mv9lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.646830 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f3643568-6595-4c73-9c83-be6e28fabe83" (UID: "f3643568-6595-4c73-9c83-be6e28fabe83"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.695411 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3643568-6595-4c73-9c83-be6e28fabe83" (UID: "f3643568-6595-4c73-9c83-be6e28fabe83"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.708572 4815 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3643568-6595-4c73-9c83-be6e28fabe83-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.708636 4815 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.708651 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.708665 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mv9lz\" (UniqueName: \"kubernetes.io/projected/f3643568-6595-4c73-9c83-be6e28fabe83-kube-api-access-mv9lz\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.708678 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.708689 4815 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3643568-6595-4c73-9c83-be6e28fabe83-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.710956 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-config-data" (OuterVolumeSpecName: "config-data") pod "f3643568-6595-4c73-9c83-be6e28fabe83" (UID: "f3643568-6595-4c73-9c83-be6e28fabe83"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.729681 4815 generic.go:334] "Generic (PLEG): container finished" podID="f3643568-6595-4c73-9c83-be6e28fabe83" containerID="df5eca668b03228f87a515ecd41160a6cf652cf136a652b575cfa3c9868694e9" exitCode=0 Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.729727 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3643568-6595-4c73-9c83-be6e28fabe83","Type":"ContainerDied","Data":"df5eca668b03228f87a515ecd41160a6cf652cf136a652b575cfa3c9868694e9"} Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.729756 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3643568-6595-4c73-9c83-be6e28fabe83","Type":"ContainerDied","Data":"7f8fd4524137989ea9b83636c13dfa58a0ca76969ae71f4c03aab4111d99265f"} Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.729778 4815 scope.go:117] "RemoveContainer" containerID="47cb76120db6d64fb0e087e27d5cbbb53191df2db146504b0d521e66c55ec08f" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.729910 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.800459 4815 scope.go:117] "RemoveContainer" containerID="431e99a3dca863f8134146b4183f20a39dd63d734e655e385c0aceb2f42d819f" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.805907 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.810148 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3643568-6595-4c73-9c83-be6e28fabe83-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.816122 4815 scope.go:117] "RemoveContainer" containerID="df5eca668b03228f87a515ecd41160a6cf652cf136a652b575cfa3c9868694e9" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.822594 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.843418 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.843440 4815 scope.go:117] "RemoveContainer" containerID="dc7434d19b0960010335e00c9621ca0164b153f5cd542c93cb16504e4883f085" Feb 25 13:38:34 crc kubenswrapper[4815]: E0225 13:38:34.844206 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3643568-6595-4c73-9c83-be6e28fabe83" containerName="ceilometer-central-agent" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.844228 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3643568-6595-4c73-9c83-be6e28fabe83" containerName="ceilometer-central-agent" Feb 25 13:38:34 crc kubenswrapper[4815]: E0225 13:38:34.844246 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3643568-6595-4c73-9c83-be6e28fabe83" containerName="sg-core" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.844257 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3643568-6595-4c73-9c83-be6e28fabe83" containerName="sg-core" Feb 25 13:38:34 crc kubenswrapper[4815]: E0225 13:38:34.844280 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3643568-6595-4c73-9c83-be6e28fabe83" containerName="proxy-httpd" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.844287 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3643568-6595-4c73-9c83-be6e28fabe83" containerName="proxy-httpd" Feb 25 13:38:34 crc kubenswrapper[4815]: E0225 13:38:34.844315 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3643568-6595-4c73-9c83-be6e28fabe83" containerName="ceilometer-notification-agent" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.844322 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3643568-6595-4c73-9c83-be6e28fabe83" containerName="ceilometer-notification-agent" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.845278 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3643568-6595-4c73-9c83-be6e28fabe83" containerName="sg-core" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.845300 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3643568-6595-4c73-9c83-be6e28fabe83" containerName="ceilometer-central-agent" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.845314 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3643568-6595-4c73-9c83-be6e28fabe83" containerName="proxy-httpd" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.845334 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3643568-6595-4c73-9c83-be6e28fabe83" containerName="ceilometer-notification-agent" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.848185 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.852556 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.852907 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.853290 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.898846 4815 scope.go:117] "RemoveContainer" containerID="47cb76120db6d64fb0e087e27d5cbbb53191df2db146504b0d521e66c55ec08f" Feb 25 13:38:34 crc kubenswrapper[4815]: E0225 13:38:34.901223 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47cb76120db6d64fb0e087e27d5cbbb53191df2db146504b0d521e66c55ec08f\": container with ID starting with 47cb76120db6d64fb0e087e27d5cbbb53191df2db146504b0d521e66c55ec08f not found: ID does not exist" containerID="47cb76120db6d64fb0e087e27d5cbbb53191df2db146504b0d521e66c55ec08f" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.901283 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47cb76120db6d64fb0e087e27d5cbbb53191df2db146504b0d521e66c55ec08f"} err="failed to get container status \"47cb76120db6d64fb0e087e27d5cbbb53191df2db146504b0d521e66c55ec08f\": rpc error: code = NotFound desc = could not find container \"47cb76120db6d64fb0e087e27d5cbbb53191df2db146504b0d521e66c55ec08f\": container with ID starting with 47cb76120db6d64fb0e087e27d5cbbb53191df2db146504b0d521e66c55ec08f not found: ID does not exist" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.901316 4815 scope.go:117] "RemoveContainer" containerID="431e99a3dca863f8134146b4183f20a39dd63d734e655e385c0aceb2f42d819f" Feb 25 13:38:34 crc kubenswrapper[4815]: E0225 13:38:34.905817 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"431e99a3dca863f8134146b4183f20a39dd63d734e655e385c0aceb2f42d819f\": container with ID starting with 431e99a3dca863f8134146b4183f20a39dd63d734e655e385c0aceb2f42d819f not found: ID does not exist" containerID="431e99a3dca863f8134146b4183f20a39dd63d734e655e385c0aceb2f42d819f" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.905861 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"431e99a3dca863f8134146b4183f20a39dd63d734e655e385c0aceb2f42d819f"} err="failed to get container status \"431e99a3dca863f8134146b4183f20a39dd63d734e655e385c0aceb2f42d819f\": rpc error: code = NotFound desc = could not find container \"431e99a3dca863f8134146b4183f20a39dd63d734e655e385c0aceb2f42d819f\": container with ID starting with 431e99a3dca863f8134146b4183f20a39dd63d734e655e385c0aceb2f42d819f not found: ID does not exist" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.905887 4815 scope.go:117] "RemoveContainer" containerID="df5eca668b03228f87a515ecd41160a6cf652cf136a652b575cfa3c9868694e9" Feb 25 13:38:34 crc kubenswrapper[4815]: E0225 13:38:34.909727 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df5eca668b03228f87a515ecd41160a6cf652cf136a652b575cfa3c9868694e9\": container with ID starting with df5eca668b03228f87a515ecd41160a6cf652cf136a652b575cfa3c9868694e9 not found: ID does not exist" containerID="df5eca668b03228f87a515ecd41160a6cf652cf136a652b575cfa3c9868694e9" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.909759 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df5eca668b03228f87a515ecd41160a6cf652cf136a652b575cfa3c9868694e9"} err="failed to get container status \"df5eca668b03228f87a515ecd41160a6cf652cf136a652b575cfa3c9868694e9\": rpc error: code = NotFound desc = could not find container \"df5eca668b03228f87a515ecd41160a6cf652cf136a652b575cfa3c9868694e9\": container with ID starting with df5eca668b03228f87a515ecd41160a6cf652cf136a652b575cfa3c9868694e9 not found: ID does not exist" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.909780 4815 scope.go:117] "RemoveContainer" containerID="dc7434d19b0960010335e00c9621ca0164b153f5cd542c93cb16504e4883f085" Feb 25 13:38:34 crc kubenswrapper[4815]: E0225 13:38:34.910050 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc7434d19b0960010335e00c9621ca0164b153f5cd542c93cb16504e4883f085\": container with ID starting with dc7434d19b0960010335e00c9621ca0164b153f5cd542c93cb16504e4883f085 not found: ID does not exist" containerID="dc7434d19b0960010335e00c9621ca0164b153f5cd542c93cb16504e4883f085" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.910073 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc7434d19b0960010335e00c9621ca0164b153f5cd542c93cb16504e4883f085"} err="failed to get container status \"dc7434d19b0960010335e00c9621ca0164b153f5cd542c93cb16504e4883f085\": rpc error: code = NotFound desc = could not find container \"dc7434d19b0960010335e00c9621ca0164b153f5cd542c93cb16504e4883f085\": container with ID starting with dc7434d19b0960010335e00c9621ca0164b153f5cd542c93cb16504e4883f085 not found: ID does not exist" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.981746 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3643568-6595-4c73-9c83-be6e28fabe83" path="/var/lib/kubelet/pods/f3643568-6595-4c73-9c83-be6e28fabe83/volumes" Feb 25 13:38:34 crc kubenswrapper[4815]: I0225 13:38:34.982517 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:38:34 crc kubenswrapper[4815]: E0225 13:38:34.983381 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-kzfr9 log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.012581 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-config-data\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.012630 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.012659 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-scripts\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.012690 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-log-httpd\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.012723 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-run-httpd\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.012760 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzfr9\" (UniqueName: \"kubernetes.io/projected/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-kube-api-access-kzfr9\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.012803 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.114681 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.114781 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-config-data\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.114802 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.114830 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-scripts\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.114861 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-log-httpd\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.114895 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-run-httpd\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.114911 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzfr9\" (UniqueName: \"kubernetes.io/projected/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-kube-api-access-kzfr9\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.115386 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-log-httpd\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.115449 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-run-httpd\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.120285 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-scripts\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.122955 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.123304 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.123706 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-config-data\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.130929 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzfr9\" (UniqueName: \"kubernetes.io/projected/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-kube-api-access-kzfr9\") pod \"ceilometer-0\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.745611 4815 generic.go:334] "Generic (PLEG): container finished" podID="7bdc9071-e863-4108-9234-4f201732f435" containerID="9d5688a7d91b39b94743229dfb65ac8852aa764ec59864066e8ded936e9d1ece" exitCode=0 Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.745692 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-d4ssg" event={"ID":"7bdc9071-e863-4108-9234-4f201732f435","Type":"ContainerDied","Data":"9d5688a7d91b39b94743229dfb65ac8852aa764ec59864066e8ded936e9d1ece"} Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.746057 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.762539 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.928666 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-scripts\") pod \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.928813 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-log-httpd\") pod \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.928870 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-sg-core-conf-yaml\") pod \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.928921 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-combined-ca-bundle\") pod \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.928965 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-config-data\") pod \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.928989 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzfr9\" (UniqueName: \"kubernetes.io/projected/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-kube-api-access-kzfr9\") pod \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.929147 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-run-httpd\") pod \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\" (UID: \"fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84\") " Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.929549 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84" (UID: "fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.929755 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84" (UID: "fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.930082 4815 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.930102 4815 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.934370 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-config-data" (OuterVolumeSpecName: "config-data") pod "fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84" (UID: "fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.935832 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84" (UID: "fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.936901 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-kube-api-access-kzfr9" (OuterVolumeSpecName: "kube-api-access-kzfr9") pod "fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84" (UID: "fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84"). InnerVolumeSpecName "kube-api-access-kzfr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.936932 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-scripts" (OuterVolumeSpecName: "scripts") pod "fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84" (UID: "fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:35 crc kubenswrapper[4815]: I0225 13:38:35.937617 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84" (UID: "fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:36 crc kubenswrapper[4815]: I0225 13:38:36.032572 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:36 crc kubenswrapper[4815]: I0225 13:38:36.032631 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:36 crc kubenswrapper[4815]: I0225 13:38:36.032659 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzfr9\" (UniqueName: \"kubernetes.io/projected/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-kube-api-access-kzfr9\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:36 crc kubenswrapper[4815]: I0225 13:38:36.032690 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:36 crc kubenswrapper[4815]: I0225 13:38:36.032716 4815 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:36 crc kubenswrapper[4815]: I0225 13:38:36.755053 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:38:36 crc kubenswrapper[4815]: I0225 13:38:36.861484 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:38:36 crc kubenswrapper[4815]: I0225 13:38:36.870970 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:38:36 crc kubenswrapper[4815]: I0225 13:38:36.886970 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:38:36 crc kubenswrapper[4815]: I0225 13:38:36.889183 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:38:36 crc kubenswrapper[4815]: I0225 13:38:36.891237 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 13:38:36 crc kubenswrapper[4815]: I0225 13:38:36.893449 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:38:36 crc kubenswrapper[4815]: I0225 13:38:36.895253 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 13:38:36 crc kubenswrapper[4815]: I0225 13:38:36.944810 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84" path="/var/lib/kubelet/pods/fea5d312-ba6d-4ed0-8ed8-3552a5c8bd84/volumes" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.051347 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.051407 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/178df77c-9b85-4f6e-8c47-44f47dc60383-run-httpd\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.051457 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-scripts\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.051488 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.051574 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-config-data\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.051654 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhhwk\" (UniqueName: \"kubernetes.io/projected/178df77c-9b85-4f6e-8c47-44f47dc60383-kube-api-access-mhhwk\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.051767 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/178df77c-9b85-4f6e-8c47-44f47dc60383-log-httpd\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.133704 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-d4ssg" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.154376 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/178df77c-9b85-4f6e-8c47-44f47dc60383-log-httpd\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.155939 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/178df77c-9b85-4f6e-8c47-44f47dc60383-log-httpd\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.156431 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/178df77c-9b85-4f6e-8c47-44f47dc60383-run-httpd\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.154494 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/178df77c-9b85-4f6e-8c47-44f47dc60383-run-httpd\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.156678 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.156749 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-scripts\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.156848 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.156990 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-config-data\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.157173 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhhwk\" (UniqueName: \"kubernetes.io/projected/178df77c-9b85-4f6e-8c47-44f47dc60383-kube-api-access-mhhwk\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.166001 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.166478 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-scripts\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.167692 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.171328 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-config-data\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.183062 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhhwk\" (UniqueName: \"kubernetes.io/projected/178df77c-9b85-4f6e-8c47-44f47dc60383-kube-api-access-mhhwk\") pod \"ceilometer-0\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.214463 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.259018 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bdc9071-e863-4108-9234-4f201732f435-combined-ca-bundle\") pod \"7bdc9071-e863-4108-9234-4f201732f435\" (UID: \"7bdc9071-e863-4108-9234-4f201732f435\") " Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.259203 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7spnb\" (UniqueName: \"kubernetes.io/projected/7bdc9071-e863-4108-9234-4f201732f435-kube-api-access-7spnb\") pod \"7bdc9071-e863-4108-9234-4f201732f435\" (UID: \"7bdc9071-e863-4108-9234-4f201732f435\") " Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.259268 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bdc9071-e863-4108-9234-4f201732f435-db-sync-config-data\") pod \"7bdc9071-e863-4108-9234-4f201732f435\" (UID: \"7bdc9071-e863-4108-9234-4f201732f435\") " Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.264951 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bdc9071-e863-4108-9234-4f201732f435-kube-api-access-7spnb" (OuterVolumeSpecName: "kube-api-access-7spnb") pod "7bdc9071-e863-4108-9234-4f201732f435" (UID: "7bdc9071-e863-4108-9234-4f201732f435"). InnerVolumeSpecName "kube-api-access-7spnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.265233 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bdc9071-e863-4108-9234-4f201732f435-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7bdc9071-e863-4108-9234-4f201732f435" (UID: "7bdc9071-e863-4108-9234-4f201732f435"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.287745 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bdc9071-e863-4108-9234-4f201732f435-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7bdc9071-e863-4108-9234-4f201732f435" (UID: "7bdc9071-e863-4108-9234-4f201732f435"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.363039 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bdc9071-e863-4108-9234-4f201732f435-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.363323 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7spnb\" (UniqueName: \"kubernetes.io/projected/7bdc9071-e863-4108-9234-4f201732f435-kube-api-access-7spnb\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.363341 4815 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7bdc9071-e863-4108-9234-4f201732f435-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.697291 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:38:37 crc kubenswrapper[4815]: W0225 13:38:37.703290 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod178df77c_9b85_4f6e_8c47_44f47dc60383.slice/crio-92e2f09c8e6644a7dc04b6f9a552463c154899a3cbf4b5e6a8c762cc72c72457 WatchSource:0}: Error finding container 92e2f09c8e6644a7dc04b6f9a552463c154899a3cbf4b5e6a8c762cc72c72457: Status 404 returned error can't find the container with id 92e2f09c8e6644a7dc04b6f9a552463c154899a3cbf4b5e6a8c762cc72c72457 Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.765552 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"178df77c-9b85-4f6e-8c47-44f47dc60383","Type":"ContainerStarted","Data":"92e2f09c8e6644a7dc04b6f9a552463c154899a3cbf4b5e6a8c762cc72c72457"} Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.767260 4815 generic.go:334] "Generic (PLEG): container finished" podID="04e3d721-2179-4457-8cc6-72af3b3f9257" containerID="4ce2c3c9b64d5f047c35dc1e7ac70a87bf7be1bc271d659eef2e7ae14c2e1a8e" exitCode=0 Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.767342 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-576gs" event={"ID":"04e3d721-2179-4457-8cc6-72af3b3f9257","Type":"ContainerDied","Data":"4ce2c3c9b64d5f047c35dc1e7ac70a87bf7be1bc271d659eef2e7ae14c2e1a8e"} Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.769392 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-d4ssg" event={"ID":"7bdc9071-e863-4108-9234-4f201732f435","Type":"ContainerDied","Data":"fbbd68a725326ce247e9bca7810965261c130ba37fa09d8cc74505e052af8d96"} Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.769432 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbbd68a725326ce247e9bca7810965261c130ba37fa09d8cc74505e052af8d96" Feb 25 13:38:37 crc kubenswrapper[4815]: I0225 13:38:37.769460 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-d4ssg" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.127349 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-77f7d79d79-6znrw"] Feb 25 13:38:38 crc kubenswrapper[4815]: E0225 13:38:38.128192 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bdc9071-e863-4108-9234-4f201732f435" containerName="barbican-db-sync" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.128214 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bdc9071-e863-4108-9234-4f201732f435" containerName="barbican-db-sync" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.128535 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bdc9071-e863-4108-9234-4f201732f435" containerName="barbican-db-sync" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.129677 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.136618 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mkf9f" Feb 25 13:38:38 crc kubenswrapper[4815]: W0225 13:38:38.136639 4815 reflector.go:561] object-"openstack"/"barbican-worker-config-data": failed to list *v1.Secret: secrets "barbican-worker-config-data" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Feb 25 13:38:38 crc kubenswrapper[4815]: E0225 13:38:38.136676 4815 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"barbican-worker-config-data\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"barbican-worker-config-data\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.136792 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.144067 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-b4dcd45c8-7j8v9"] Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.157831 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.179157 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-77f7d79d79-6znrw"] Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.204949 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/877d6ac2-0a5e-43bd-b387-0a2ba928b54f-logs\") pod \"barbican-keystone-listener-b4dcd45c8-7j8v9\" (UID: \"877d6ac2-0a5e-43bd-b387-0a2ba928b54f\") " pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.205005 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df0b5667-833f-40d6-8ef4-5c988adae0ba-config-data-custom\") pod \"barbican-worker-77f7d79d79-6znrw\" (UID: \"df0b5667-833f-40d6-8ef4-5c988adae0ba\") " pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.205046 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvwc4\" (UniqueName: \"kubernetes.io/projected/877d6ac2-0a5e-43bd-b387-0a2ba928b54f-kube-api-access-wvwc4\") pod \"barbican-keystone-listener-b4dcd45c8-7j8v9\" (UID: \"877d6ac2-0a5e-43bd-b387-0a2ba928b54f\") " pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.205067 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/877d6ac2-0a5e-43bd-b387-0a2ba928b54f-config-data-custom\") pod \"barbican-keystone-listener-b4dcd45c8-7j8v9\" (UID: \"877d6ac2-0a5e-43bd-b387-0a2ba928b54f\") " pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.205092 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s62mt\" (UniqueName: \"kubernetes.io/projected/df0b5667-833f-40d6-8ef4-5c988adae0ba-kube-api-access-s62mt\") pod \"barbican-worker-77f7d79d79-6znrw\" (UID: \"df0b5667-833f-40d6-8ef4-5c988adae0ba\") " pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.205114 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/877d6ac2-0a5e-43bd-b387-0a2ba928b54f-combined-ca-bundle\") pod \"barbican-keystone-listener-b4dcd45c8-7j8v9\" (UID: \"877d6ac2-0a5e-43bd-b387-0a2ba928b54f\") " pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.205133 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df0b5667-833f-40d6-8ef4-5c988adae0ba-config-data\") pod \"barbican-worker-77f7d79d79-6znrw\" (UID: \"df0b5667-833f-40d6-8ef4-5c988adae0ba\") " pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.205162 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/877d6ac2-0a5e-43bd-b387-0a2ba928b54f-config-data\") pod \"barbican-keystone-listener-b4dcd45c8-7j8v9\" (UID: \"877d6ac2-0a5e-43bd-b387-0a2ba928b54f\") " pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.205190 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df0b5667-833f-40d6-8ef4-5c988adae0ba-combined-ca-bundle\") pod \"barbican-worker-77f7d79d79-6znrw\" (UID: \"df0b5667-833f-40d6-8ef4-5c988adae0ba\") " pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.205211 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df0b5667-833f-40d6-8ef4-5c988adae0ba-logs\") pod \"barbican-worker-77f7d79d79-6znrw\" (UID: \"df0b5667-833f-40d6-8ef4-5c988adae0ba\") " pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.205602 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.266249 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-b4dcd45c8-7j8v9"] Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.310322 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/877d6ac2-0a5e-43bd-b387-0a2ba928b54f-config-data\") pod \"barbican-keystone-listener-b4dcd45c8-7j8v9\" (UID: \"877d6ac2-0a5e-43bd-b387-0a2ba928b54f\") " pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.310574 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df0b5667-833f-40d6-8ef4-5c988adae0ba-combined-ca-bundle\") pod \"barbican-worker-77f7d79d79-6znrw\" (UID: \"df0b5667-833f-40d6-8ef4-5c988adae0ba\") " pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.310660 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df0b5667-833f-40d6-8ef4-5c988adae0ba-logs\") pod \"barbican-worker-77f7d79d79-6znrw\" (UID: \"df0b5667-833f-40d6-8ef4-5c988adae0ba\") " pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.310877 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/877d6ac2-0a5e-43bd-b387-0a2ba928b54f-logs\") pod \"barbican-keystone-listener-b4dcd45c8-7j8v9\" (UID: \"877d6ac2-0a5e-43bd-b387-0a2ba928b54f\") " pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.310991 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df0b5667-833f-40d6-8ef4-5c988adae0ba-config-data-custom\") pod \"barbican-worker-77f7d79d79-6znrw\" (UID: \"df0b5667-833f-40d6-8ef4-5c988adae0ba\") " pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.311138 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvwc4\" (UniqueName: \"kubernetes.io/projected/877d6ac2-0a5e-43bd-b387-0a2ba928b54f-kube-api-access-wvwc4\") pod \"barbican-keystone-listener-b4dcd45c8-7j8v9\" (UID: \"877d6ac2-0a5e-43bd-b387-0a2ba928b54f\") " pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.311189 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/877d6ac2-0a5e-43bd-b387-0a2ba928b54f-config-data-custom\") pod \"barbican-keystone-listener-b4dcd45c8-7j8v9\" (UID: \"877d6ac2-0a5e-43bd-b387-0a2ba928b54f\") " pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.311262 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s62mt\" (UniqueName: \"kubernetes.io/projected/df0b5667-833f-40d6-8ef4-5c988adae0ba-kube-api-access-s62mt\") pod \"barbican-worker-77f7d79d79-6znrw\" (UID: \"df0b5667-833f-40d6-8ef4-5c988adae0ba\") " pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.311328 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/877d6ac2-0a5e-43bd-b387-0a2ba928b54f-combined-ca-bundle\") pod \"barbican-keystone-listener-b4dcd45c8-7j8v9\" (UID: \"877d6ac2-0a5e-43bd-b387-0a2ba928b54f\") " pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.311392 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df0b5667-833f-40d6-8ef4-5c988adae0ba-config-data\") pod \"barbican-worker-77f7d79d79-6znrw\" (UID: \"df0b5667-833f-40d6-8ef4-5c988adae0ba\") " pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.312043 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/877d6ac2-0a5e-43bd-b387-0a2ba928b54f-logs\") pod \"barbican-keystone-listener-b4dcd45c8-7j8v9\" (UID: \"877d6ac2-0a5e-43bd-b387-0a2ba928b54f\") " pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.318323 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df0b5667-833f-40d6-8ef4-5c988adae0ba-logs\") pod \"barbican-worker-77f7d79d79-6znrw\" (UID: \"df0b5667-833f-40d6-8ef4-5c988adae0ba\") " pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.324501 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-zgkl2"] Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.327062 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df0b5667-833f-40d6-8ef4-5c988adae0ba-combined-ca-bundle\") pod \"barbican-worker-77f7d79d79-6znrw\" (UID: \"df0b5667-833f-40d6-8ef4-5c988adae0ba\") " pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.327190 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/877d6ac2-0a5e-43bd-b387-0a2ba928b54f-config-data-custom\") pod \"barbican-keystone-listener-b4dcd45c8-7j8v9\" (UID: \"877d6ac2-0a5e-43bd-b387-0a2ba928b54f\") " pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.327334 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df0b5667-833f-40d6-8ef4-5c988adae0ba-config-data\") pod \"barbican-worker-77f7d79d79-6znrw\" (UID: \"df0b5667-833f-40d6-8ef4-5c988adae0ba\") " pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.330996 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/877d6ac2-0a5e-43bd-b387-0a2ba928b54f-config-data\") pod \"barbican-keystone-listener-b4dcd45c8-7j8v9\" (UID: \"877d6ac2-0a5e-43bd-b387-0a2ba928b54f\") " pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.342215 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/877d6ac2-0a5e-43bd-b387-0a2ba928b54f-combined-ca-bundle\") pod \"barbican-keystone-listener-b4dcd45c8-7j8v9\" (UID: \"877d6ac2-0a5e-43bd-b387-0a2ba928b54f\") " pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.345095 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s62mt\" (UniqueName: \"kubernetes.io/projected/df0b5667-833f-40d6-8ef4-5c988adae0ba-kube-api-access-s62mt\") pod \"barbican-worker-77f7d79d79-6znrw\" (UID: \"df0b5667-833f-40d6-8ef4-5c988adae0ba\") " pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.351082 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvwc4\" (UniqueName: \"kubernetes.io/projected/877d6ac2-0a5e-43bd-b387-0a2ba928b54f-kube-api-access-wvwc4\") pod \"barbican-keystone-listener-b4dcd45c8-7j8v9\" (UID: \"877d6ac2-0a5e-43bd-b387-0a2ba928b54f\") " pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.351241 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-zgkl2"] Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.351355 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.375352 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-b86d988f8-9zdkl"] Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.380985 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.384543 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.424416 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-b86d988f8-9zdkl"] Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.514495 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-config\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.514780 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-combined-ca-bundle\") pod \"barbican-api-b86d988f8-9zdkl\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.514828 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-config-data\") pod \"barbican-api-b86d988f8-9zdkl\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.514870 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7sdl\" (UniqueName: \"kubernetes.io/projected/8114370d-9aee-4979-9445-2f0f57e805cf-kube-api-access-t7sdl\") pod \"barbican-api-b86d988f8-9zdkl\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.514899 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8114370d-9aee-4979-9445-2f0f57e805cf-logs\") pod \"barbican-api-b86d988f8-9zdkl\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.515503 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-dns-svc\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.515609 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-config-data-custom\") pod \"barbican-api-b86d988f8-9zdkl\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.516042 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.516101 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.516137 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k629j\" (UniqueName: \"kubernetes.io/projected/1afce2e2-a088-4005-b394-56f943e88509-kube-api-access-k629j\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.516167 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.562117 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.617703 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7sdl\" (UniqueName: \"kubernetes.io/projected/8114370d-9aee-4979-9445-2f0f57e805cf-kube-api-access-t7sdl\") pod \"barbican-api-b86d988f8-9zdkl\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.617774 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8114370d-9aee-4979-9445-2f0f57e805cf-logs\") pod \"barbican-api-b86d988f8-9zdkl\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.617805 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-dns-svc\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.617864 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-config-data-custom\") pod \"barbican-api-b86d988f8-9zdkl\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.617910 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.617947 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.617977 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k629j\" (UniqueName: \"kubernetes.io/projected/1afce2e2-a088-4005-b394-56f943e88509-kube-api-access-k629j\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.618001 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.618040 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-config\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.618077 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-combined-ca-bundle\") pod \"barbican-api-b86d988f8-9zdkl\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.618104 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-config-data\") pod \"barbican-api-b86d988f8-9zdkl\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.620383 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.620957 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.621301 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-config\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.621542 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8114370d-9aee-4979-9445-2f0f57e805cf-logs\") pod \"barbican-api-b86d988f8-9zdkl\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.622878 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.623034 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-dns-svc\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.625986 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-combined-ca-bundle\") pod \"barbican-api-b86d988f8-9zdkl\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.626094 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-config-data-custom\") pod \"barbican-api-b86d988f8-9zdkl\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.627164 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-config-data\") pod \"barbican-api-b86d988f8-9zdkl\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.641378 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k629j\" (UniqueName: \"kubernetes.io/projected/1afce2e2-a088-4005-b394-56f943e88509-kube-api-access-k629j\") pod \"dnsmasq-dns-688c87cc99-zgkl2\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.641998 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7sdl\" (UniqueName: \"kubernetes.io/projected/8114370d-9aee-4979-9445-2f0f57e805cf-kube-api-access-t7sdl\") pod \"barbican-api-b86d988f8-9zdkl\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.726538 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.744166 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:38 crc kubenswrapper[4815]: I0225 13:38:38.783745 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"178df77c-9b85-4f6e-8c47-44f47dc60383","Type":"ContainerStarted","Data":"dfe32d3f63f2e76c406dc47bcfa57eb2c27bc3a91d00b5753fa4909f18a475b3"} Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.015612 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-b4dcd45c8-7j8v9"] Feb 25 13:38:39 crc kubenswrapper[4815]: W0225 13:38:39.041906 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod877d6ac2_0a5e_43bd_b387_0a2ba928b54f.slice/crio-5af995164a7ef1c93c47b713fcf040dca0593d31bfc9f46b504cc4c638ab8a8d WatchSource:0}: Error finding container 5af995164a7ef1c93c47b713fcf040dca0593d31bfc9f46b504cc4c638ab8a8d: Status 404 returned error can't find the container with id 5af995164a7ef1c93c47b713fcf040dca0593d31bfc9f46b504cc4c638ab8a8d Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.168747 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-zgkl2"] Feb 25 13:38:39 crc kubenswrapper[4815]: W0225 13:38:39.182052 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1afce2e2_a088_4005_b394_56f943e88509.slice/crio-ca732d85fb57b810b0d4106544383b1a66f29addc7a30b88521ab30a4236c01f WatchSource:0}: Error finding container ca732d85fb57b810b0d4106544383b1a66f29addc7a30b88521ab30a4236c01f: Status 404 returned error can't find the container with id ca732d85fb57b810b0d4106544383b1a66f29addc7a30b88521ab30a4236c01f Feb 25 13:38:39 crc kubenswrapper[4815]: E0225 13:38:39.319755 4815 secret.go:188] Couldn't get secret openstack/barbican-worker-config-data: failed to sync secret cache: timed out waiting for the condition Feb 25 13:38:39 crc kubenswrapper[4815]: E0225 13:38:39.320205 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/df0b5667-833f-40d6-8ef4-5c988adae0ba-config-data-custom podName:df0b5667-833f-40d6-8ef4-5c988adae0ba nodeName:}" failed. No retries permitted until 2026-02-25 13:38:39.820169037 +0000 UTC m=+1077.621267091 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/df0b5667-833f-40d6-8ef4-5c988adae0ba-config-data-custom") pod "barbican-worker-77f7d79d79-6znrw" (UID: "df0b5667-833f-40d6-8ef4-5c988adae0ba") : failed to sync secret cache: timed out waiting for the condition Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.328534 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.402459 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-b86d988f8-9zdkl"] Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.426715 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-576gs" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.461014 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/04e3d721-2179-4457-8cc6-72af3b3f9257-etc-machine-id\") pod \"04e3d721-2179-4457-8cc6-72af3b3f9257\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.461206 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9dbn\" (UniqueName: \"kubernetes.io/projected/04e3d721-2179-4457-8cc6-72af3b3f9257-kube-api-access-j9dbn\") pod \"04e3d721-2179-4457-8cc6-72af3b3f9257\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.461242 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-scripts\") pod \"04e3d721-2179-4457-8cc6-72af3b3f9257\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.461272 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-config-data\") pod \"04e3d721-2179-4457-8cc6-72af3b3f9257\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.461333 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-combined-ca-bundle\") pod \"04e3d721-2179-4457-8cc6-72af3b3f9257\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.461126 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04e3d721-2179-4457-8cc6-72af3b3f9257-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "04e3d721-2179-4457-8cc6-72af3b3f9257" (UID: "04e3d721-2179-4457-8cc6-72af3b3f9257"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.461990 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-db-sync-config-data\") pod \"04e3d721-2179-4457-8cc6-72af3b3f9257\" (UID: \"04e3d721-2179-4457-8cc6-72af3b3f9257\") " Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.462619 4815 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/04e3d721-2179-4457-8cc6-72af3b3f9257-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.466938 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-scripts" (OuterVolumeSpecName: "scripts") pod "04e3d721-2179-4457-8cc6-72af3b3f9257" (UID: "04e3d721-2179-4457-8cc6-72af3b3f9257"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.468020 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "04e3d721-2179-4457-8cc6-72af3b3f9257" (UID: "04e3d721-2179-4457-8cc6-72af3b3f9257"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.469727 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04e3d721-2179-4457-8cc6-72af3b3f9257-kube-api-access-j9dbn" (OuterVolumeSpecName: "kube-api-access-j9dbn") pod "04e3d721-2179-4457-8cc6-72af3b3f9257" (UID: "04e3d721-2179-4457-8cc6-72af3b3f9257"). InnerVolumeSpecName "kube-api-access-j9dbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.525812 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04e3d721-2179-4457-8cc6-72af3b3f9257" (UID: "04e3d721-2179-4457-8cc6-72af3b3f9257"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.564415 4815 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.564441 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9dbn\" (UniqueName: \"kubernetes.io/projected/04e3d721-2179-4457-8cc6-72af3b3f9257-kube-api-access-j9dbn\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.564455 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.564465 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.570077 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-config-data" (OuterVolumeSpecName: "config-data") pod "04e3d721-2179-4457-8cc6-72af3b3f9257" (UID: "04e3d721-2179-4457-8cc6-72af3b3f9257"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.666082 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04e3d721-2179-4457-8cc6-72af3b3f9257-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.797676 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-576gs" event={"ID":"04e3d721-2179-4457-8cc6-72af3b3f9257","Type":"ContainerDied","Data":"24b242a8edf5fb48d59efcd722a53de16f270444b84375f434536b74320b3ed7"} Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.798012 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24b242a8edf5fb48d59efcd722a53de16f270444b84375f434536b74320b3ed7" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.798084 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-576gs" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.802749 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" event={"ID":"877d6ac2-0a5e-43bd-b387-0a2ba928b54f","Type":"ContainerStarted","Data":"5af995164a7ef1c93c47b713fcf040dca0593d31bfc9f46b504cc4c638ab8a8d"} Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.804264 4815 generic.go:334] "Generic (PLEG): container finished" podID="1afce2e2-a088-4005-b394-56f943e88509" containerID="49e4879504474db3de5c95f66ece19dc51dbfa2cd0e15b5a4b48fd5ddab36ba4" exitCode=0 Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.804307 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" event={"ID":"1afce2e2-a088-4005-b394-56f943e88509","Type":"ContainerDied","Data":"49e4879504474db3de5c95f66ece19dc51dbfa2cd0e15b5a4b48fd5ddab36ba4"} Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.804322 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" event={"ID":"1afce2e2-a088-4005-b394-56f943e88509","Type":"ContainerStarted","Data":"ca732d85fb57b810b0d4106544383b1a66f29addc7a30b88521ab30a4236c01f"} Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.816596 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b86d988f8-9zdkl" event={"ID":"8114370d-9aee-4979-9445-2f0f57e805cf","Type":"ContainerStarted","Data":"7989163bcdd6a0940634a28e03ec0e2678fb563e455feefb7f59ab23861a70f3"} Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.816727 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b86d988f8-9zdkl" event={"ID":"8114370d-9aee-4979-9445-2f0f57e805cf","Type":"ContainerStarted","Data":"54356262d93d48a51a0e4589f1246a5b05e27e0966783a0604b95de28ec29155"} Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.816750 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b86d988f8-9zdkl" event={"ID":"8114370d-9aee-4979-9445-2f0f57e805cf","Type":"ContainerStarted","Data":"7d9efe2a61521b7fe2b1800cd02c7d39f0b57a0806ac39e11741992588bad282"} Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.816770 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.816867 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.843931 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"178df77c-9b85-4f6e-8c47-44f47dc60383","Type":"ContainerStarted","Data":"d7e37413d82a64dba8f1ff0c987f3cc5793ccc301f538991657226e2a5025542"} Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.844026 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"178df77c-9b85-4f6e-8c47-44f47dc60383","Type":"ContainerStarted","Data":"0e3f2c6c27b9e6a889af97e0163bfa7460186ec10b5b75dc96cf6c63752c31b3"} Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.863969 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-b86d988f8-9zdkl" podStartSLOduration=1.8639490520000002 podStartE2EDuration="1.863949052s" podCreationTimestamp="2026-02-25 13:38:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:39.843425852 +0000 UTC m=+1077.644523946" watchObservedRunningTime="2026-02-25 13:38:39.863949052 +0000 UTC m=+1077.665047106" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.868691 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df0b5667-833f-40d6-8ef4-5c988adae0ba-config-data-custom\") pod \"barbican-worker-77f7d79d79-6znrw\" (UID: \"df0b5667-833f-40d6-8ef4-5c988adae0ba\") " pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:39 crc kubenswrapper[4815]: I0225 13:38:39.873984 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df0b5667-833f-40d6-8ef4-5c988adae0ba-config-data-custom\") pod \"barbican-worker-77f7d79d79-6znrw\" (UID: \"df0b5667-833f-40d6-8ef4-5c988adae0ba\") " pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.035392 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-77f7d79d79-6znrw" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.089219 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 13:38:40 crc kubenswrapper[4815]: E0225 13:38:40.089598 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04e3d721-2179-4457-8cc6-72af3b3f9257" containerName="cinder-db-sync" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.089615 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="04e3d721-2179-4457-8cc6-72af3b3f9257" containerName="cinder-db-sync" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.089808 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="04e3d721-2179-4457-8cc6-72af3b3f9257" containerName="cinder-db-sync" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.090669 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.111590 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.111659 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-zjkm2" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.111849 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.112158 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.116054 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.152417 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-zgkl2"] Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.174066 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.174117 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c7055a7-b147-4629-906c-ecfcc968733b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.174165 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.174212 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlnjh\" (UniqueName: \"kubernetes.io/projected/9c7055a7-b147-4629-906c-ecfcc968733b-kube-api-access-jlnjh\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.174246 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-config-data\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.174267 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-scripts\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.195359 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2wmhn"] Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.197658 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.233959 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2wmhn"] Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.275675 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-config-data\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.275725 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-scripts\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.275774 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.275807 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.275823 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-config\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.275852 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c7055a7-b147-4629-906c-ecfcc968733b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.275871 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.275894 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr7gg\" (UniqueName: \"kubernetes.io/projected/2e02f6a9-4025-45bc-a523-17b0c0758268-kube-api-access-cr7gg\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.275911 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.275944 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.275964 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.276009 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlnjh\" (UniqueName: \"kubernetes.io/projected/9c7055a7-b147-4629-906c-ecfcc968733b-kube-api-access-jlnjh\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.277631 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c7055a7-b147-4629-906c-ecfcc968733b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.282038 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-scripts\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.285667 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-config-data\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.286030 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.286433 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.354038 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlnjh\" (UniqueName: \"kubernetes.io/projected/9c7055a7-b147-4629-906c-ecfcc968733b-kube-api-access-jlnjh\") pod \"cinder-scheduler-0\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.375598 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.377631 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.379041 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr7gg\" (UniqueName: \"kubernetes.io/projected/2e02f6a9-4025-45bc-a523-17b0c0758268-kube-api-access-cr7gg\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.379088 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.379148 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.379330 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.379374 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-config\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.379418 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.380259 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.380524 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.381120 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.386287 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.386353 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.387655 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-config\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.406966 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr7gg\" (UniqueName: \"kubernetes.io/projected/2e02f6a9-4025-45bc-a523-17b0c0758268-kube-api-access-cr7gg\") pod \"dnsmasq-dns-6bb4fc677f-2wmhn\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.418073 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.478312 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.480605 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-config-data\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.480659 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-scripts\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.480824 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3e66813f-2f10-4c69-8b54-f6c678f75f57-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.480928 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-config-data-custom\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.481065 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.481102 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tktxd\" (UniqueName: \"kubernetes.io/projected/3e66813f-2f10-4c69-8b54-f6c678f75f57-kube-api-access-tktxd\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.481145 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e66813f-2f10-4c69-8b54-f6c678f75f57-logs\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.550800 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.582521 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-config-data\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.582570 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-scripts\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.582622 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3e66813f-2f10-4c69-8b54-f6c678f75f57-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.582656 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-config-data-custom\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.582707 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.582729 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tktxd\" (UniqueName: \"kubernetes.io/projected/3e66813f-2f10-4c69-8b54-f6c678f75f57-kube-api-access-tktxd\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.582766 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e66813f-2f10-4c69-8b54-f6c678f75f57-logs\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.583266 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e66813f-2f10-4c69-8b54-f6c678f75f57-logs\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.587017 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3e66813f-2f10-4c69-8b54-f6c678f75f57-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.587601 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-config-data-custom\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.589911 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-config-data\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.590167 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.593664 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-scripts\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.601911 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tktxd\" (UniqueName: \"kubernetes.io/projected/3e66813f-2f10-4c69-8b54-f6c678f75f57-kube-api-access-tktxd\") pod \"cinder-api-0\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.776956 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.884080 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" event={"ID":"1afce2e2-a088-4005-b394-56f943e88509","Type":"ContainerStarted","Data":"ae5272d39d061e74b93c883542f8e9088397b480fe8fa5440ba38e24904a1709"} Feb 25 13:38:40 crc kubenswrapper[4815]: I0225 13:38:40.912679 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" podStartSLOduration=2.9126617660000003 podStartE2EDuration="2.912661766s" podCreationTimestamp="2026-02-25 13:38:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:40.906678981 +0000 UTC m=+1078.707777035" watchObservedRunningTime="2026-02-25 13:38:40.912661766 +0000 UTC m=+1078.713759820" Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.284949 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-77f7d79d79-6znrw"] Feb 25 13:38:41 crc kubenswrapper[4815]: W0225 13:38:41.291542 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf0b5667_833f_40d6_8ef4_5c988adae0ba.slice/crio-f13109c5852fc5629e492a8afbf622b9bf740477e45dade1427cd7685dc0fa54 WatchSource:0}: Error finding container f13109c5852fc5629e492a8afbf622b9bf740477e45dade1427cd7685dc0fa54: Status 404 returned error can't find the container with id f13109c5852fc5629e492a8afbf622b9bf740477e45dade1427cd7685dc0fa54 Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.383423 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2wmhn"] Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.392639 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.399004 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 25 13:38:41 crc kubenswrapper[4815]: W0225 13:38:41.400650 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e66813f_2f10_4c69_8b54_f6c678f75f57.slice/crio-94d8f489b8dddd22966cc8237efe7dd00cacd68a55eb1c20b075a9b5834156c9 WatchSource:0}: Error finding container 94d8f489b8dddd22966cc8237efe7dd00cacd68a55eb1c20b075a9b5834156c9: Status 404 returned error can't find the container with id 94d8f489b8dddd22966cc8237efe7dd00cacd68a55eb1c20b075a9b5834156c9 Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.896563 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" event={"ID":"877d6ac2-0a5e-43bd-b387-0a2ba928b54f","Type":"ContainerStarted","Data":"ad3ef25c76b5f9bb2c45f687a543ca0120d1937cac4954a35360c3392b6507cc"} Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.896857 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" event={"ID":"877d6ac2-0a5e-43bd-b387-0a2ba928b54f","Type":"ContainerStarted","Data":"b021573e53df7b03e4ce8605abb523159ddb32d1664201f2c044790bee2fecce"} Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.898743 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3e66813f-2f10-4c69-8b54-f6c678f75f57","Type":"ContainerStarted","Data":"94d8f489b8dddd22966cc8237efe7dd00cacd68a55eb1c20b075a9b5834156c9"} Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.901378 4815 generic.go:334] "Generic (PLEG): container finished" podID="2e02f6a9-4025-45bc-a523-17b0c0758268" containerID="466bc2996fc37fe1a22e3361b2eeffbadf14ff5bfc43b0515011ba2561ad00c6" exitCode=0 Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.901437 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" event={"ID":"2e02f6a9-4025-45bc-a523-17b0c0758268","Type":"ContainerDied","Data":"466bc2996fc37fe1a22e3361b2eeffbadf14ff5bfc43b0515011ba2561ad00c6"} Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.901463 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" event={"ID":"2e02f6a9-4025-45bc-a523-17b0c0758268","Type":"ContainerStarted","Data":"8b65475db2ee56caa78e1c32b26aba990137f8fe1a5a4450720600f3f92735a2"} Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.906982 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"178df77c-9b85-4f6e-8c47-44f47dc60383","Type":"ContainerStarted","Data":"64fa5eafb58ae0796dbda8aa80d9b18239391d9bff07332f0ee6268798453139"} Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.907893 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.909768 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77f7d79d79-6znrw" event={"ID":"df0b5667-833f-40d6-8ef4-5c988adae0ba","Type":"ContainerStarted","Data":"f13109c5852fc5629e492a8afbf622b9bf740477e45dade1427cd7685dc0fa54"} Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.913407 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9c7055a7-b147-4629-906c-ecfcc968733b","Type":"ContainerStarted","Data":"3f291da4e56c94bd06610767bfbc97a029f86de1d9a87fe1ebda5882d60f83e4"} Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.913560 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" podUID="1afce2e2-a088-4005-b394-56f943e88509" containerName="dnsmasq-dns" containerID="cri-o://ae5272d39d061e74b93c883542f8e9088397b480fe8fa5440ba38e24904a1709" gracePeriod=10 Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.913700 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.931942 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-b4dcd45c8-7j8v9" podStartSLOduration=2.216175567 podStartE2EDuration="3.931923999s" podCreationTimestamp="2026-02-25 13:38:38 +0000 UTC" firstStartedPulling="2026-02-25 13:38:39.047529978 +0000 UTC m=+1076.848628022" lastFinishedPulling="2026-02-25 13:38:40.7632784 +0000 UTC m=+1078.564376454" observedRunningTime="2026-02-25 13:38:41.918482907 +0000 UTC m=+1079.719580971" watchObservedRunningTime="2026-02-25 13:38:41.931923999 +0000 UTC m=+1079.733022053" Feb 25 13:38:41 crc kubenswrapper[4815]: I0225 13:38:41.948952 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.279377515 podStartE2EDuration="5.948930347s" podCreationTimestamp="2026-02-25 13:38:36 +0000 UTC" firstStartedPulling="2026-02-25 13:38:37.70582493 +0000 UTC m=+1075.506922984" lastFinishedPulling="2026-02-25 13:38:41.375377762 +0000 UTC m=+1079.176475816" observedRunningTime="2026-02-25 13:38:41.9476574 +0000 UTC m=+1079.748755474" watchObservedRunningTime="2026-02-25 13:38:41.948930347 +0000 UTC m=+1079.750028411" Feb 25 13:38:42 crc kubenswrapper[4815]: I0225 13:38:42.850327 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:42 crc kubenswrapper[4815]: I0225 13:38:42.927736 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" event={"ID":"2e02f6a9-4025-45bc-a523-17b0c0758268","Type":"ContainerStarted","Data":"f2077ba5aa6728327f15f64309d6fecfec0e0dd7553e042a47e119e5d1a5a998"} Feb 25 13:38:42 crc kubenswrapper[4815]: I0225 13:38:42.929961 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:42 crc kubenswrapper[4815]: I0225 13:38:42.932705 4815 generic.go:334] "Generic (PLEG): container finished" podID="1afce2e2-a088-4005-b394-56f943e88509" containerID="ae5272d39d061e74b93c883542f8e9088397b480fe8fa5440ba38e24904a1709" exitCode=0 Feb 25 13:38:42 crc kubenswrapper[4815]: I0225 13:38:42.932760 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" Feb 25 13:38:42 crc kubenswrapper[4815]: I0225 13:38:42.932762 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" event={"ID":"1afce2e2-a088-4005-b394-56f943e88509","Type":"ContainerDied","Data":"ae5272d39d061e74b93c883542f8e9088397b480fe8fa5440ba38e24904a1709"} Feb 25 13:38:42 crc kubenswrapper[4815]: I0225 13:38:42.932892 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-zgkl2" event={"ID":"1afce2e2-a088-4005-b394-56f943e88509","Type":"ContainerDied","Data":"ca732d85fb57b810b0d4106544383b1a66f29addc7a30b88521ab30a4236c01f"} Feb 25 13:38:42 crc kubenswrapper[4815]: I0225 13:38:42.932929 4815 scope.go:117] "RemoveContainer" containerID="ae5272d39d061e74b93c883542f8e9088397b480fe8fa5440ba38e24904a1709" Feb 25 13:38:42 crc kubenswrapper[4815]: I0225 13:38:42.936913 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-ovsdbserver-nb\") pod \"1afce2e2-a088-4005-b394-56f943e88509\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " Feb 25 13:38:42 crc kubenswrapper[4815]: I0225 13:38:42.937036 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-dns-swift-storage-0\") pod \"1afce2e2-a088-4005-b394-56f943e88509\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " Feb 25 13:38:42 crc kubenswrapper[4815]: I0225 13:38:42.937123 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k629j\" (UniqueName: \"kubernetes.io/projected/1afce2e2-a088-4005-b394-56f943e88509-kube-api-access-k629j\") pod \"1afce2e2-a088-4005-b394-56f943e88509\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " Feb 25 13:38:42 crc kubenswrapper[4815]: I0225 13:38:42.937159 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-config\") pod \"1afce2e2-a088-4005-b394-56f943e88509\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " Feb 25 13:38:42 crc kubenswrapper[4815]: I0225 13:38:42.937231 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-ovsdbserver-sb\") pod \"1afce2e2-a088-4005-b394-56f943e88509\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " Feb 25 13:38:42 crc kubenswrapper[4815]: I0225 13:38:42.937262 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-dns-svc\") pod \"1afce2e2-a088-4005-b394-56f943e88509\" (UID: \"1afce2e2-a088-4005-b394-56f943e88509\") " Feb 25 13:38:42 crc kubenswrapper[4815]: I0225 13:38:42.972316 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1afce2e2-a088-4005-b394-56f943e88509-kube-api-access-k629j" (OuterVolumeSpecName: "kube-api-access-k629j") pod "1afce2e2-a088-4005-b394-56f943e88509" (UID: "1afce2e2-a088-4005-b394-56f943e88509"). InnerVolumeSpecName "kube-api-access-k629j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:42 crc kubenswrapper[4815]: I0225 13:38:42.981381 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" podStartSLOduration=2.981361265 podStartE2EDuration="2.981361265s" podCreationTimestamp="2026-02-25 13:38:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:42.954112498 +0000 UTC m=+1080.755210552" watchObservedRunningTime="2026-02-25 13:38:42.981361265 +0000 UTC m=+1080.782459329" Feb 25 13:38:42 crc kubenswrapper[4815]: I0225 13:38:42.997045 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3e66813f-2f10-4c69-8b54-f6c678f75f57","Type":"ContainerStarted","Data":"b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467"} Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.058134 4815 scope.go:117] "RemoveContainer" containerID="49e4879504474db3de5c95f66ece19dc51dbfa2cd0e15b5a4b48fd5ddab36ba4" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.061266 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k629j\" (UniqueName: \"kubernetes.io/projected/1afce2e2-a088-4005-b394-56f943e88509-kube-api-access-k629j\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.136768 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1afce2e2-a088-4005-b394-56f943e88509" (UID: "1afce2e2-a088-4005-b394-56f943e88509"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.151732 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1afce2e2-a088-4005-b394-56f943e88509" (UID: "1afce2e2-a088-4005-b394-56f943e88509"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.151946 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.158420 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-config" (OuterVolumeSpecName: "config") pod "1afce2e2-a088-4005-b394-56f943e88509" (UID: "1afce2e2-a088-4005-b394-56f943e88509"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.158832 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1afce2e2-a088-4005-b394-56f943e88509" (UID: "1afce2e2-a088-4005-b394-56f943e88509"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.163687 4815 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.163724 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.163735 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.163746 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.168102 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1afce2e2-a088-4005-b394-56f943e88509" (UID: "1afce2e2-a088-4005-b394-56f943e88509"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.265337 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1afce2e2-a088-4005-b394-56f943e88509-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.277272 4815 scope.go:117] "RemoveContainer" containerID="ae5272d39d061e74b93c883542f8e9088397b480fe8fa5440ba38e24904a1709" Feb 25 13:38:43 crc kubenswrapper[4815]: E0225 13:38:43.282679 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae5272d39d061e74b93c883542f8e9088397b480fe8fa5440ba38e24904a1709\": container with ID starting with ae5272d39d061e74b93c883542f8e9088397b480fe8fa5440ba38e24904a1709 not found: ID does not exist" containerID="ae5272d39d061e74b93c883542f8e9088397b480fe8fa5440ba38e24904a1709" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.282731 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae5272d39d061e74b93c883542f8e9088397b480fe8fa5440ba38e24904a1709"} err="failed to get container status \"ae5272d39d061e74b93c883542f8e9088397b480fe8fa5440ba38e24904a1709\": rpc error: code = NotFound desc = could not find container \"ae5272d39d061e74b93c883542f8e9088397b480fe8fa5440ba38e24904a1709\": container with ID starting with ae5272d39d061e74b93c883542f8e9088397b480fe8fa5440ba38e24904a1709 not found: ID does not exist" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.282767 4815 scope.go:117] "RemoveContainer" containerID="49e4879504474db3de5c95f66ece19dc51dbfa2cd0e15b5a4b48fd5ddab36ba4" Feb 25 13:38:43 crc kubenswrapper[4815]: E0225 13:38:43.283204 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49e4879504474db3de5c95f66ece19dc51dbfa2cd0e15b5a4b48fd5ddab36ba4\": container with ID starting with 49e4879504474db3de5c95f66ece19dc51dbfa2cd0e15b5a4b48fd5ddab36ba4 not found: ID does not exist" containerID="49e4879504474db3de5c95f66ece19dc51dbfa2cd0e15b5a4b48fd5ddab36ba4" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.283233 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49e4879504474db3de5c95f66ece19dc51dbfa2cd0e15b5a4b48fd5ddab36ba4"} err="failed to get container status \"49e4879504474db3de5c95f66ece19dc51dbfa2cd0e15b5a4b48fd5ddab36ba4\": rpc error: code = NotFound desc = could not find container \"49e4879504474db3de5c95f66ece19dc51dbfa2cd0e15b5a4b48fd5ddab36ba4\": container with ID starting with 49e4879504474db3de5c95f66ece19dc51dbfa2cd0e15b5a4b48fd5ddab36ba4 not found: ID does not exist" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.324900 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.380915 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-zgkl2"] Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.389308 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-zgkl2"] Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.601253 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-645c577cc7-vpgnt"] Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.602318 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-645c577cc7-vpgnt" podUID="145d5de3-c56d-4209-b3b7-348f2f6d2365" containerName="neutron-httpd" containerID="cri-o://97585b874331e8484826968cbdaea6a4929b28c335a1547f8af1086da98b9aad" gracePeriod=30 Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.601816 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-645c577cc7-vpgnt" podUID="145d5de3-c56d-4209-b3b7-348f2f6d2365" containerName="neutron-api" containerID="cri-o://4c38128e88e041b8ab297bea5065cb3651ef99685d28363708fceec5f74ca125" gracePeriod=30 Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.617350 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-645c577cc7-vpgnt" podUID="145d5de3-c56d-4209-b3b7-348f2f6d2365" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.158:9696/\": read tcp 10.217.0.2:41834->10.217.0.158:9696: read: connection reset by peer" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.620481 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-68dffb47c7-mfx8z"] Feb 25 13:38:43 crc kubenswrapper[4815]: E0225 13:38:43.623922 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1afce2e2-a088-4005-b394-56f943e88509" containerName="init" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.623949 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1afce2e2-a088-4005-b394-56f943e88509" containerName="init" Feb 25 13:38:43 crc kubenswrapper[4815]: E0225 13:38:43.623962 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1afce2e2-a088-4005-b394-56f943e88509" containerName="dnsmasq-dns" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.623997 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1afce2e2-a088-4005-b394-56f943e88509" containerName="dnsmasq-dns" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.624196 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1afce2e2-a088-4005-b394-56f943e88509" containerName="dnsmasq-dns" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.625548 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.638839 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-68dffb47c7-mfx8z"] Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.776499 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-combined-ca-bundle\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.776809 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwgcd\" (UniqueName: \"kubernetes.io/projected/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-kube-api-access-fwgcd\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.776850 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-ovndb-tls-certs\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.776907 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-internal-tls-certs\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.776936 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-public-tls-certs\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.776980 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-httpd-config\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.777052 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-config\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.878569 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwgcd\" (UniqueName: \"kubernetes.io/projected/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-kube-api-access-fwgcd\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.878617 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-ovndb-tls-certs\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.878641 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-internal-tls-certs\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.878662 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-public-tls-certs\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.878688 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-httpd-config\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.878712 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-config\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.878786 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-combined-ca-bundle\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.882552 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-internal-tls-certs\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.882944 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-combined-ca-bundle\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.883820 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-ovndb-tls-certs\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.883888 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-config\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.884626 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-public-tls-certs\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.884710 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-httpd-config\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.894131 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwgcd\" (UniqueName: \"kubernetes.io/projected/2ba9ece6-3c27-41f3-ab94-f4f73b12a265-kube-api-access-fwgcd\") pod \"neutron-68dffb47c7-mfx8z\" (UID: \"2ba9ece6-3c27-41f3-ab94-f4f73b12a265\") " pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.947358 4815 scope.go:117] "RemoveContainer" containerID="54171993046dd1f6abdbe71ffe2e0962df40a595a20f014f4d7593c4403cfcb9" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.959239 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3e66813f-2f10-4c69-8b54-f6c678f75f57","Type":"ContainerStarted","Data":"f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749"} Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.959311 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3e66813f-2f10-4c69-8b54-f6c678f75f57" containerName="cinder-api-log" containerID="cri-o://b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467" gracePeriod=30 Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.959444 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3e66813f-2f10-4c69-8b54-f6c678f75f57" containerName="cinder-api" containerID="cri-o://f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749" gracePeriod=30 Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.959582 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.965999 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77f7d79d79-6znrw" event={"ID":"df0b5667-833f-40d6-8ef4-5c988adae0ba","Type":"ContainerStarted","Data":"048eb954db28057c8e7780b36367efc52b8f57c312f55103bcd1352c1c4c0f80"} Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.966035 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77f7d79d79-6znrw" event={"ID":"df0b5667-833f-40d6-8ef4-5c988adae0ba","Type":"ContainerStarted","Data":"d05f3e8b9c6d003aab6b0aeea84d67398a69aa345d834c877d57c410480b5ba8"} Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.974368 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9c7055a7-b147-4629-906c-ecfcc968733b","Type":"ContainerStarted","Data":"653acdd019d486b2d24004094ba6855b8e3416fd5cce26d5e7b10c7ae8129191"} Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.983180 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.983163168 podStartE2EDuration="3.983163168s" podCreationTimestamp="2026-02-25 13:38:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:43.973141265 +0000 UTC m=+1081.774239339" watchObservedRunningTime="2026-02-25 13:38:43.983163168 +0000 UTC m=+1081.784261222" Feb 25 13:38:43 crc kubenswrapper[4815]: I0225 13:38:43.983971 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.017963 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-77f7d79d79-6znrw" podStartSLOduration=4.536223974 podStartE2EDuration="6.017943434s" podCreationTimestamp="2026-02-25 13:38:38 +0000 UTC" firstStartedPulling="2026-02-25 13:38:41.295121986 +0000 UTC m=+1079.096220040" lastFinishedPulling="2026-02-25 13:38:42.776841446 +0000 UTC m=+1080.577939500" observedRunningTime="2026-02-25 13:38:43.987941678 +0000 UTC m=+1081.789039732" watchObservedRunningTime="2026-02-25 13:38:44.017943434 +0000 UTC m=+1081.819041488" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.592754 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-68dffb47c7-mfx8z"] Feb 25 13:38:44 crc kubenswrapper[4815]: W0225 13:38:44.608620 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ba9ece6_3c27_41f3_ab94_f4f73b12a265.slice/crio-eeed2d38d32e06a2d0711347b30930ad36403c53fe32366c220fd39b18b6550a WatchSource:0}: Error finding container eeed2d38d32e06a2d0711347b30930ad36403c53fe32366c220fd39b18b6550a: Status 404 returned error can't find the container with id eeed2d38d32e06a2d0711347b30930ad36403c53fe32366c220fd39b18b6550a Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.650917 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.801916 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3e66813f-2f10-4c69-8b54-f6c678f75f57-etc-machine-id\") pod \"3e66813f-2f10-4c69-8b54-f6c678f75f57\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.801988 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e66813f-2f10-4c69-8b54-f6c678f75f57-logs\") pod \"3e66813f-2f10-4c69-8b54-f6c678f75f57\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.802078 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-combined-ca-bundle\") pod \"3e66813f-2f10-4c69-8b54-f6c678f75f57\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.802163 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-scripts\") pod \"3e66813f-2f10-4c69-8b54-f6c678f75f57\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.802230 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-config-data-custom\") pod \"3e66813f-2f10-4c69-8b54-f6c678f75f57\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.802246 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-config-data\") pod \"3e66813f-2f10-4c69-8b54-f6c678f75f57\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.802291 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tktxd\" (UniqueName: \"kubernetes.io/projected/3e66813f-2f10-4c69-8b54-f6c678f75f57-kube-api-access-tktxd\") pod \"3e66813f-2f10-4c69-8b54-f6c678f75f57\" (UID: \"3e66813f-2f10-4c69-8b54-f6c678f75f57\") " Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.802986 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e66813f-2f10-4c69-8b54-f6c678f75f57-logs" (OuterVolumeSpecName: "logs") pod "3e66813f-2f10-4c69-8b54-f6c678f75f57" (UID: "3e66813f-2f10-4c69-8b54-f6c678f75f57"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.803083 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e66813f-2f10-4c69-8b54-f6c678f75f57-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3e66813f-2f10-4c69-8b54-f6c678f75f57" (UID: "3e66813f-2f10-4c69-8b54-f6c678f75f57"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.809100 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.812597 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-scripts" (OuterVolumeSpecName: "scripts") pod "3e66813f-2f10-4c69-8b54-f6c678f75f57" (UID: "3e66813f-2f10-4c69-8b54-f6c678f75f57"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.813837 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e66813f-2f10-4c69-8b54-f6c678f75f57-kube-api-access-tktxd" (OuterVolumeSpecName: "kube-api-access-tktxd") pod "3e66813f-2f10-4c69-8b54-f6c678f75f57" (UID: "3e66813f-2f10-4c69-8b54-f6c678f75f57"). InnerVolumeSpecName "kube-api-access-tktxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.814524 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3e66813f-2f10-4c69-8b54-f6c678f75f57" (UID: "3e66813f-2f10-4c69-8b54-f6c678f75f57"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.857691 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e66813f-2f10-4c69-8b54-f6c678f75f57" (UID: "3e66813f-2f10-4c69-8b54-f6c678f75f57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.888312 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-config-data" (OuterVolumeSpecName: "config-data") pod "3e66813f-2f10-4c69-8b54-f6c678f75f57" (UID: "3e66813f-2f10-4c69-8b54-f6c678f75f57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.904185 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9c18809-9f62-4b7e-8103-df382e9b4fa6-logs\") pod \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.904262 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d9c18809-9f62-4b7e-8103-df382e9b4fa6-config-data\") pod \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.904433 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d9c18809-9f62-4b7e-8103-df382e9b4fa6-horizon-secret-key\") pod \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.904486 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d9c18809-9f62-4b7e-8103-df382e9b4fa6-scripts\") pod \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.904502 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmp45\" (UniqueName: \"kubernetes.io/projected/d9c18809-9f62-4b7e-8103-df382e9b4fa6-kube-api-access-cmp45\") pod \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\" (UID: \"d9c18809-9f62-4b7e-8103-df382e9b4fa6\") " Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.905101 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.905118 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.905127 4815 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.905135 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e66813f-2f10-4c69-8b54-f6c678f75f57-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.905144 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tktxd\" (UniqueName: \"kubernetes.io/projected/3e66813f-2f10-4c69-8b54-f6c678f75f57-kube-api-access-tktxd\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.905154 4815 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3e66813f-2f10-4c69-8b54-f6c678f75f57-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.905162 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e66813f-2f10-4c69-8b54-f6c678f75f57-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.907761 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9c18809-9f62-4b7e-8103-df382e9b4fa6-logs" (OuterVolumeSpecName: "logs") pod "d9c18809-9f62-4b7e-8103-df382e9b4fa6" (UID: "d9c18809-9f62-4b7e-8103-df382e9b4fa6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.952201 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9c18809-9f62-4b7e-8103-df382e9b4fa6-config-data" (OuterVolumeSpecName: "config-data") pod "d9c18809-9f62-4b7e-8103-df382e9b4fa6" (UID: "d9c18809-9f62-4b7e-8103-df382e9b4fa6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.952387 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9c18809-9f62-4b7e-8103-df382e9b4fa6-kube-api-access-cmp45" (OuterVolumeSpecName: "kube-api-access-cmp45") pod "d9c18809-9f62-4b7e-8103-df382e9b4fa6" (UID: "d9c18809-9f62-4b7e-8103-df382e9b4fa6"). InnerVolumeSpecName "kube-api-access-cmp45". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.954656 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9c18809-9f62-4b7e-8103-df382e9b4fa6-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d9c18809-9f62-4b7e-8103-df382e9b4fa6" (UID: "d9c18809-9f62-4b7e-8103-df382e9b4fa6"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:44 crc kubenswrapper[4815]: I0225 13:38:44.984040 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1afce2e2-a088-4005-b394-56f943e88509" path="/var/lib/kubelet/pods/1afce2e2-a088-4005-b394-56f943e88509/volumes" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.013751 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9c18809-9f62-4b7e-8103-df382e9b4fa6-scripts" (OuterVolumeSpecName: "scripts") pod "d9c18809-9f62-4b7e-8103-df382e9b4fa6" (UID: "d9c18809-9f62-4b7e-8103-df382e9b4fa6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.015016 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d9c18809-9f62-4b7e-8103-df382e9b4fa6-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.015099 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmp45\" (UniqueName: \"kubernetes.io/projected/d9c18809-9f62-4b7e-8103-df382e9b4fa6-kube-api-access-cmp45\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.015155 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9c18809-9f62-4b7e-8103-df382e9b4fa6-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.015205 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d9c18809-9f62-4b7e-8103-df382e9b4fa6-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.015257 4815 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d9c18809-9f62-4b7e-8103-df382e9b4fa6-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.024014 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68dffb47c7-mfx8z" event={"ID":"2ba9ece6-3c27-41f3-ab94-f4f73b12a265","Type":"ContainerStarted","Data":"c040cbbf19b9cea5e9a385d17ca77a9ee05818dc31d03a7de7e05cb25e520298"} Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.024369 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68dffb47c7-mfx8z" event={"ID":"2ba9ece6-3c27-41f3-ab94-f4f73b12a265","Type":"ContainerStarted","Data":"eeed2d38d32e06a2d0711347b30930ad36403c53fe32366c220fd39b18b6550a"} Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.041629 4815 generic.go:334] "Generic (PLEG): container finished" podID="3e66813f-2f10-4c69-8b54-f6c678f75f57" containerID="f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749" exitCode=0 Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.041657 4815 generic.go:334] "Generic (PLEG): container finished" podID="3e66813f-2f10-4c69-8b54-f6c678f75f57" containerID="b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467" exitCode=143 Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.042083 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.042720 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3e66813f-2f10-4c69-8b54-f6c678f75f57","Type":"ContainerDied","Data":"f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749"} Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.042774 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3e66813f-2f10-4c69-8b54-f6c678f75f57","Type":"ContainerDied","Data":"b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467"} Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.042786 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3e66813f-2f10-4c69-8b54-f6c678f75f57","Type":"ContainerDied","Data":"94d8f489b8dddd22966cc8237efe7dd00cacd68a55eb1c20b075a9b5834156c9"} Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.042806 4815 scope.go:117] "RemoveContainer" containerID="f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.051667 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7fc6fdc978-tzrxw"] Feb 25 13:38:45 crc kubenswrapper[4815]: E0225 13:38:45.052315 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9c18809-9f62-4b7e-8103-df382e9b4fa6" containerName="horizon-log" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.052441 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9c18809-9f62-4b7e-8103-df382e9b4fa6" containerName="horizon-log" Feb 25 13:38:45 crc kubenswrapper[4815]: E0225 13:38:45.052473 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e66813f-2f10-4c69-8b54-f6c678f75f57" containerName="cinder-api-log" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.052481 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e66813f-2f10-4c69-8b54-f6c678f75f57" containerName="cinder-api-log" Feb 25 13:38:45 crc kubenswrapper[4815]: E0225 13:38:45.052538 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9c18809-9f62-4b7e-8103-df382e9b4fa6" containerName="horizon" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.052547 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9c18809-9f62-4b7e-8103-df382e9b4fa6" containerName="horizon" Feb 25 13:38:45 crc kubenswrapper[4815]: E0225 13:38:45.052560 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e66813f-2f10-4c69-8b54-f6c678f75f57" containerName="cinder-api" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.052567 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e66813f-2f10-4c69-8b54-f6c678f75f57" containerName="cinder-api" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.052893 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9c18809-9f62-4b7e-8103-df382e9b4fa6" containerName="horizon" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.052960 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e66813f-2f10-4c69-8b54-f6c678f75f57" containerName="cinder-api" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.052986 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e66813f-2f10-4c69-8b54-f6c678f75f57" containerName="cinder-api-log" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.053003 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9c18809-9f62-4b7e-8103-df382e9b4fa6" containerName="horizon-log" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.055255 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.055433 4815 generic.go:334] "Generic (PLEG): container finished" podID="145d5de3-c56d-4209-b3b7-348f2f6d2365" containerID="97585b874331e8484826968cbdaea6a4929b28c335a1547f8af1086da98b9aad" exitCode=0 Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.055567 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-645c577cc7-vpgnt" event={"ID":"145d5de3-c56d-4209-b3b7-348f2f6d2365","Type":"ContainerDied","Data":"97585b874331e8484826968cbdaea6a4929b28c335a1547f8af1086da98b9aad"} Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.070699 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.071322 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.073936 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9c7055a7-b147-4629-906c-ecfcc968733b","Type":"ContainerStarted","Data":"09af70168a6ece2b5d62aa4e754f7f7c5e2cb722a29afe4fb88f1a10f341719e"} Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.079123 4815 generic.go:334] "Generic (PLEG): container finished" podID="d9c18809-9f62-4b7e-8103-df382e9b4fa6" containerID="3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f" exitCode=137 Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.079160 4815 generic.go:334] "Generic (PLEG): container finished" podID="d9c18809-9f62-4b7e-8103-df382e9b4fa6" containerID="46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102" exitCode=137 Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.080108 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c455f8ff5-tkdz5" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.096184 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c455f8ff5-tkdz5" event={"ID":"d9c18809-9f62-4b7e-8103-df382e9b4fa6","Type":"ContainerDied","Data":"3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f"} Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.096259 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7fc6fdc978-tzrxw"] Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.096288 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c455f8ff5-tkdz5" event={"ID":"d9c18809-9f62-4b7e-8103-df382e9b4fa6","Type":"ContainerDied","Data":"46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102"} Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.096300 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c455f8ff5-tkdz5" event={"ID":"d9c18809-9f62-4b7e-8103-df382e9b4fa6","Type":"ContainerDied","Data":"50acff9ff504baddfe9194e3d3802a75cd3bfd040c6efa1bf77167cea43ec545"} Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.109184 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.118559 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.127583 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.129176 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.135580 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.135758 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.135862 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.177288 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.184891 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.813945852 podStartE2EDuration="5.184872114s" podCreationTimestamp="2026-02-25 13:38:40 +0000 UTC" firstStartedPulling="2026-02-25 13:38:41.399076085 +0000 UTC m=+1079.200174129" lastFinishedPulling="2026-02-25 13:38:42.770002337 +0000 UTC m=+1080.571100391" observedRunningTime="2026-02-25 13:38:45.148797489 +0000 UTC m=+1082.949895543" watchObservedRunningTime="2026-02-25 13:38:45.184872114 +0000 UTC m=+1082.985970168" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.216863 4815 scope.go:117] "RemoveContainer" containerID="b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.221299 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd01efd5-a5c6-449d-b539-f31454945163-logs\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.221337 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwnng\" (UniqueName: \"kubernetes.io/projected/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-kube-api-access-cwnng\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.221361 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-logs\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.221395 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd01efd5-a5c6-449d-b539-f31454945163-internal-tls-certs\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.221909 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-config-data\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.221944 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd01efd5-a5c6-449d-b539-f31454945163-config-data\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.221998 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd01efd5-a5c6-449d-b539-f31454945163-config-data-custom\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.222091 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.222200 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-scripts\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.222897 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd01efd5-a5c6-449d-b539-f31454945163-combined-ca-bundle\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.222976 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzmjh\" (UniqueName: \"kubernetes.io/projected/cd01efd5-a5c6-449d-b539-f31454945163-kube-api-access-qzmjh\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.223094 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.223147 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-config-data-custom\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.223204 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.223244 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd01efd5-a5c6-449d-b539-f31454945163-public-tls-certs\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.223320 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.236147 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c455f8ff5-tkdz5"] Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.245760 4815 scope.go:117] "RemoveContainer" containerID="f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749" Feb 25 13:38:45 crc kubenswrapper[4815]: E0225 13:38:45.246174 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749\": container with ID starting with f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749 not found: ID does not exist" containerID="f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.246201 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749"} err="failed to get container status \"f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749\": rpc error: code = NotFound desc = could not find container \"f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749\": container with ID starting with f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749 not found: ID does not exist" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.246241 4815 scope.go:117] "RemoveContainer" containerID="b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467" Feb 25 13:38:45 crc kubenswrapper[4815]: E0225 13:38:45.246399 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467\": container with ID starting with b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467 not found: ID does not exist" containerID="b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.246431 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467"} err="failed to get container status \"b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467\": rpc error: code = NotFound desc = could not find container \"b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467\": container with ID starting with b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467 not found: ID does not exist" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.246443 4815 scope.go:117] "RemoveContainer" containerID="f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.246672 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749"} err="failed to get container status \"f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749\": rpc error: code = NotFound desc = could not find container \"f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749\": container with ID starting with f664298182c6d40a8069e5ffefbc249347a23b19cdf297c73ab0c06238dd4749 not found: ID does not exist" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.246688 4815 scope.go:117] "RemoveContainer" containerID="b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.246849 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467"} err="failed to get container status \"b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467\": rpc error: code = NotFound desc = could not find container \"b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467\": container with ID starting with b20210ba1776bf38794710c6be65a9ad28247330fd944ea6894b7bc692545467 not found: ID does not exist" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.246863 4815 scope.go:117] "RemoveContainer" containerID="3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.255489 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-c455f8ff5-tkdz5"] Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.324535 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.324583 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd01efd5-a5c6-449d-b539-f31454945163-public-tls-certs\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.324616 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.324655 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd01efd5-a5c6-449d-b539-f31454945163-logs\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.324675 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwnng\" (UniqueName: \"kubernetes.io/projected/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-kube-api-access-cwnng\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.324694 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-logs\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.324715 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd01efd5-a5c6-449d-b539-f31454945163-internal-tls-certs\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.324747 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-config-data\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.324768 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd01efd5-a5c6-449d-b539-f31454945163-config-data\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.324788 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd01efd5-a5c6-449d-b539-f31454945163-config-data-custom\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.324818 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.324840 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-scripts\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.324874 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd01efd5-a5c6-449d-b539-f31454945163-combined-ca-bundle\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.324904 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzmjh\" (UniqueName: \"kubernetes.io/projected/cd01efd5-a5c6-449d-b539-f31454945163-kube-api-access-qzmjh\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.324933 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.324952 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-config-data-custom\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.325140 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd01efd5-a5c6-449d-b539-f31454945163-logs\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.328141 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-logs\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.328190 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.330399 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd01efd5-a5c6-449d-b539-f31454945163-internal-tls-certs\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.331201 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd01efd5-a5c6-449d-b539-f31454945163-combined-ca-bundle\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.332718 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-config-data\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.332807 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd01efd5-a5c6-449d-b539-f31454945163-config-data\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.334005 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.334945 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-scripts\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.335147 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-config-data-custom\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.335188 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.335273 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.335498 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd01efd5-a5c6-449d-b539-f31454945163-public-tls-certs\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.337064 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd01efd5-a5c6-449d-b539-f31454945163-config-data-custom\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.339953 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzmjh\" (UniqueName: \"kubernetes.io/projected/cd01efd5-a5c6-449d-b539-f31454945163-kube-api-access-qzmjh\") pod \"barbican-api-7fc6fdc978-tzrxw\" (UID: \"cd01efd5-a5c6-449d-b539-f31454945163\") " pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.354082 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwnng\" (UniqueName: \"kubernetes.io/projected/fb136f80-7ca0-4315-a9f9-b36e161e6b0d-kube-api-access-cwnng\") pod \"cinder-api-0\" (UID: \"fb136f80-7ca0-4315-a9f9-b36e161e6b0d\") " pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.414224 4815 scope.go:117] "RemoveContainer" containerID="46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.463815 4815 scope.go:117] "RemoveContainer" containerID="3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f" Feb 25 13:38:45 crc kubenswrapper[4815]: E0225 13:38:45.464854 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f\": container with ID starting with 3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f not found: ID does not exist" containerID="3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.464890 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f"} err="failed to get container status \"3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f\": rpc error: code = NotFound desc = could not find container \"3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f\": container with ID starting with 3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f not found: ID does not exist" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.464916 4815 scope.go:117] "RemoveContainer" containerID="46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102" Feb 25 13:38:45 crc kubenswrapper[4815]: E0225 13:38:45.468859 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102\": container with ID starting with 46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102 not found: ID does not exist" containerID="46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.468888 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102"} err="failed to get container status \"46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102\": rpc error: code = NotFound desc = could not find container \"46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102\": container with ID starting with 46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102 not found: ID does not exist" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.468903 4815 scope.go:117] "RemoveContainer" containerID="3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.470906 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f"} err="failed to get container status \"3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f\": rpc error: code = NotFound desc = could not find container \"3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f\": container with ID starting with 3c193ff760417b0b962286ad8affda46937fa5db9fc28cab60a8b173c05ca51f not found: ID does not exist" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.470930 4815 scope.go:117] "RemoveContainer" containerID="46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.473751 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102"} err="failed to get container status \"46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102\": rpc error: code = NotFound desc = could not find container \"46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102\": container with ID starting with 46b3b90014d1f517c4c5d81fba47392b2ab672e22186ead5afb76377bfe06102 not found: ID does not exist" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.478629 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.519168 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.529098 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 13:38:45 crc kubenswrapper[4815]: I0225 13:38:45.685414 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-645c577cc7-vpgnt" podUID="145d5de3-c56d-4209-b3b7-348f2f6d2365" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.158:9696/\": dial tcp 10.217.0.158:9696: connect: connection refused" Feb 25 13:38:46 crc kubenswrapper[4815]: I0225 13:38:46.043910 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 25 13:38:46 crc kubenswrapper[4815]: W0225 13:38:46.057221 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb136f80_7ca0_4315_a9f9_b36e161e6b0d.slice/crio-8377ed6d6f06bb00bc7fd3ed769a2c68d38811bf5d5ee9ca61be7e7cffff939d WatchSource:0}: Error finding container 8377ed6d6f06bb00bc7fd3ed769a2c68d38811bf5d5ee9ca61be7e7cffff939d: Status 404 returned error can't find the container with id 8377ed6d6f06bb00bc7fd3ed769a2c68d38811bf5d5ee9ca61be7e7cffff939d Feb 25 13:38:46 crc kubenswrapper[4815]: I0225 13:38:46.112783 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fb136f80-7ca0-4315-a9f9-b36e161e6b0d","Type":"ContainerStarted","Data":"8377ed6d6f06bb00bc7fd3ed769a2c68d38811bf5d5ee9ca61be7e7cffff939d"} Feb 25 13:38:46 crc kubenswrapper[4815]: I0225 13:38:46.116453 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68dffb47c7-mfx8z" event={"ID":"2ba9ece6-3c27-41f3-ab94-f4f73b12a265","Type":"ContainerStarted","Data":"194a1d7e593333a8d8a1fa3946620ed81b2a310211ad9362b961769c99822fe1"} Feb 25 13:38:46 crc kubenswrapper[4815]: I0225 13:38:46.116661 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:38:46 crc kubenswrapper[4815]: I0225 13:38:46.163018 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-68dffb47c7-mfx8z" podStartSLOduration=3.162993245 podStartE2EDuration="3.162993245s" podCreationTimestamp="2026-02-25 13:38:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:46.142247478 +0000 UTC m=+1083.943345552" watchObservedRunningTime="2026-02-25 13:38:46.162993245 +0000 UTC m=+1083.964091299" Feb 25 13:38:46 crc kubenswrapper[4815]: I0225 13:38:46.169473 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7fc6fdc978-tzrxw"] Feb 25 13:38:46 crc kubenswrapper[4815]: I0225 13:38:46.332252 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:38:46 crc kubenswrapper[4815]: I0225 13:38:46.527274 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:38:46 crc kubenswrapper[4815]: I0225 13:38:46.949037 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e66813f-2f10-4c69-8b54-f6c678f75f57" path="/var/lib/kubelet/pods/3e66813f-2f10-4c69-8b54-f6c678f75f57/volumes" Feb 25 13:38:46 crc kubenswrapper[4815]: I0225 13:38:46.950232 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9c18809-9f62-4b7e-8103-df382e9b4fa6" path="/var/lib/kubelet/pods/d9c18809-9f62-4b7e-8103-df382e9b4fa6/volumes" Feb 25 13:38:47 crc kubenswrapper[4815]: I0225 13:38:47.141084 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fb136f80-7ca0-4315-a9f9-b36e161e6b0d","Type":"ContainerStarted","Data":"5b4afae5de22d2478b528be020f64e7b9f8371ac8edab4ac334a9e9751b22a92"} Feb 25 13:38:47 crc kubenswrapper[4815]: I0225 13:38:47.143362 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc6fdc978-tzrxw" event={"ID":"cd01efd5-a5c6-449d-b539-f31454945163","Type":"ContainerStarted","Data":"0b7e3a41a3eae1cb3a188bbe9a8e81aa2f71889f22e3228112154125869fcee1"} Feb 25 13:38:47 crc kubenswrapper[4815]: I0225 13:38:47.143394 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc6fdc978-tzrxw" event={"ID":"cd01efd5-a5c6-449d-b539-f31454945163","Type":"ContainerStarted","Data":"3346e2504d7aa78d1a9368ac530ce5c3056722b6653062a31fc3e4cddba5e265"} Feb 25 13:38:47 crc kubenswrapper[4815]: I0225 13:38:47.145618 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc6fdc978-tzrxw" event={"ID":"cd01efd5-a5c6-449d-b539-f31454945163","Type":"ContainerStarted","Data":"66e57780adb30559efa3e575025cfaefebdbfdb32e8e95a99c57b1e3ab18811d"} Feb 25 13:38:47 crc kubenswrapper[4815]: I0225 13:38:47.160139 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7fc6fdc978-tzrxw" podStartSLOduration=3.160126222 podStartE2EDuration="3.160126222s" podCreationTimestamp="2026-02-25 13:38:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:47.157716061 +0000 UTC m=+1084.958814115" watchObservedRunningTime="2026-02-25 13:38:47.160126222 +0000 UTC m=+1084.961224276" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.163036 4815 generic.go:334] "Generic (PLEG): container finished" podID="145d5de3-c56d-4209-b3b7-348f2f6d2365" containerID="4c38128e88e041b8ab297bea5065cb3651ef99685d28363708fceec5f74ca125" exitCode=0 Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.163287 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-645c577cc7-vpgnt" event={"ID":"145d5de3-c56d-4209-b3b7-348f2f6d2365","Type":"ContainerDied","Data":"4c38128e88e041b8ab297bea5065cb3651ef99685d28363708fceec5f74ca125"} Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.169223 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fb136f80-7ca0-4315-a9f9-b36e161e6b0d","Type":"ContainerStarted","Data":"4c095194fe2dc702cf5c5575bdfe16bc233724332772534525e5cfaa05331a75"} Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.170197 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.170271 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.170551 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.199902 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.199884704 podStartE2EDuration="3.199884704s" podCreationTimestamp="2026-02-25 13:38:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:48.194184577 +0000 UTC m=+1085.995282631" watchObservedRunningTime="2026-02-25 13:38:48.199884704 +0000 UTC m=+1086.000982758" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.258613 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.302311 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7657655688-k9k6t" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.366651 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6497d456d-s5g59"] Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.416224 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.603096 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-config\") pod \"145d5de3-c56d-4209-b3b7-348f2f6d2365\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.603152 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-combined-ca-bundle\") pod \"145d5de3-c56d-4209-b3b7-348f2f6d2365\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.603232 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-962nr\" (UniqueName: \"kubernetes.io/projected/145d5de3-c56d-4209-b3b7-348f2f6d2365-kube-api-access-962nr\") pod \"145d5de3-c56d-4209-b3b7-348f2f6d2365\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.603276 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-httpd-config\") pod \"145d5de3-c56d-4209-b3b7-348f2f6d2365\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.603322 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-public-tls-certs\") pod \"145d5de3-c56d-4209-b3b7-348f2f6d2365\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.603447 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-internal-tls-certs\") pod \"145d5de3-c56d-4209-b3b7-348f2f6d2365\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.603488 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-ovndb-tls-certs\") pod \"145d5de3-c56d-4209-b3b7-348f2f6d2365\" (UID: \"145d5de3-c56d-4209-b3b7-348f2f6d2365\") " Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.609707 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "145d5de3-c56d-4209-b3b7-348f2f6d2365" (UID: "145d5de3-c56d-4209-b3b7-348f2f6d2365"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.619689 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/145d5de3-c56d-4209-b3b7-348f2f6d2365-kube-api-access-962nr" (OuterVolumeSpecName: "kube-api-access-962nr") pod "145d5de3-c56d-4209-b3b7-348f2f6d2365" (UID: "145d5de3-c56d-4209-b3b7-348f2f6d2365"). InnerVolumeSpecName "kube-api-access-962nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.682015 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-config" (OuterVolumeSpecName: "config") pod "145d5de3-c56d-4209-b3b7-348f2f6d2365" (UID: "145d5de3-c56d-4209-b3b7-348f2f6d2365"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.705243 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.705273 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-962nr\" (UniqueName: \"kubernetes.io/projected/145d5de3-c56d-4209-b3b7-348f2f6d2365-kube-api-access-962nr\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.705283 4815 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.707539 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "145d5de3-c56d-4209-b3b7-348f2f6d2365" (UID: "145d5de3-c56d-4209-b3b7-348f2f6d2365"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.722781 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "145d5de3-c56d-4209-b3b7-348f2f6d2365" (UID: "145d5de3-c56d-4209-b3b7-348f2f6d2365"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.727958 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "145d5de3-c56d-4209-b3b7-348f2f6d2365" (UID: "145d5de3-c56d-4209-b3b7-348f2f6d2365"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.745420 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "145d5de3-c56d-4209-b3b7-348f2f6d2365" (UID: "145d5de3-c56d-4209-b3b7-348f2f6d2365"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.806393 4815 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.806422 4815 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.806430 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:48 crc kubenswrapper[4815]: I0225 13:38:48.806438 4815 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/145d5de3-c56d-4209-b3b7-348f2f6d2365-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:49 crc kubenswrapper[4815]: I0225 13:38:49.195009 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-645c577cc7-vpgnt" event={"ID":"145d5de3-c56d-4209-b3b7-348f2f6d2365","Type":"ContainerDied","Data":"ee07f1d14da2fcfdb7fcb74abe66b41f3176ad1f70fc65e0cc057f8220e04c03"} Feb 25 13:38:49 crc kubenswrapper[4815]: I0225 13:38:49.195071 4815 scope.go:117] "RemoveContainer" containerID="97585b874331e8484826968cbdaea6a4929b28c335a1547f8af1086da98b9aad" Feb 25 13:38:49 crc kubenswrapper[4815]: I0225 13:38:49.195646 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-645c577cc7-vpgnt" Feb 25 13:38:49 crc kubenswrapper[4815]: I0225 13:38:49.195894 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6497d456d-s5g59" podUID="5d076fbb-4499-456f-82ee-211a8909fc99" containerName="horizon-log" containerID="cri-o://37b004a1e59b2810a4a4e2a9e741191118c1925114ad66b49a3a6e92a2d9e11e" gracePeriod=30 Feb 25 13:38:49 crc kubenswrapper[4815]: I0225 13:38:49.196861 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6497d456d-s5g59" podUID="5d076fbb-4499-456f-82ee-211a8909fc99" containerName="horizon" containerID="cri-o://9fba6bb1e21ffc87dd876586c19b77148678fd408e930bd6088e1ed1c188d47d" gracePeriod=30 Feb 25 13:38:49 crc kubenswrapper[4815]: I0225 13:38:49.240575 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-645c577cc7-vpgnt"] Feb 25 13:38:49 crc kubenswrapper[4815]: I0225 13:38:49.241584 4815 scope.go:117] "RemoveContainer" containerID="4c38128e88e041b8ab297bea5065cb3651ef99685d28363708fceec5f74ca125" Feb 25 13:38:49 crc kubenswrapper[4815]: I0225 13:38:49.247604 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-645c577cc7-vpgnt"] Feb 25 13:38:50 crc kubenswrapper[4815]: I0225 13:38:50.175957 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:50 crc kubenswrapper[4815]: I0225 13:38:50.371565 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:38:50 crc kubenswrapper[4815]: I0225 13:38:50.552791 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:38:50 crc kubenswrapper[4815]: I0225 13:38:50.611808 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-gggp2"] Feb 25 13:38:50 crc kubenswrapper[4815]: I0225 13:38:50.612081 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" podUID="ea8d6d97-73b0-4db9-8712-31eb46ef0105" containerName="dnsmasq-dns" containerID="cri-o://0c4fecacd29410071ab5edabaf794d418b4b3db1af96dc7c45b4fdc9554e3452" gracePeriod=10 Feb 25 13:38:50 crc kubenswrapper[4815]: I0225 13:38:50.756716 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 25 13:38:50 crc kubenswrapper[4815]: I0225 13:38:50.796007 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 13:38:50 crc kubenswrapper[4815]: I0225 13:38:50.961581 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="145d5de3-c56d-4209-b3b7-348f2f6d2365" path="/var/lib/kubelet/pods/145d5de3-c56d-4209-b3b7-348f2f6d2365/volumes" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.168702 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.219886 4815 generic.go:334] "Generic (PLEG): container finished" podID="ea8d6d97-73b0-4db9-8712-31eb46ef0105" containerID="0c4fecacd29410071ab5edabaf794d418b4b3db1af96dc7c45b4fdc9554e3452" exitCode=0 Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.220285 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="9c7055a7-b147-4629-906c-ecfcc968733b" containerName="cinder-scheduler" containerID="cri-o://653acdd019d486b2d24004094ba6855b8e3416fd5cce26d5e7b10c7ae8129191" gracePeriod=30 Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.220648 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.221111 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" event={"ID":"ea8d6d97-73b0-4db9-8712-31eb46ef0105","Type":"ContainerDied","Data":"0c4fecacd29410071ab5edabaf794d418b4b3db1af96dc7c45b4fdc9554e3452"} Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.221191 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-gggp2" event={"ID":"ea8d6d97-73b0-4db9-8712-31eb46ef0105","Type":"ContainerDied","Data":"612731f99c90e554c44c1b407cdc68e3b8d6131a290dcd415a0a14c2859bcac9"} Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.221255 4815 scope.go:117] "RemoveContainer" containerID="0c4fecacd29410071ab5edabaf794d418b4b3db1af96dc7c45b4fdc9554e3452" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.221638 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="9c7055a7-b147-4629-906c-ecfcc968733b" containerName="probe" containerID="cri-o://09af70168a6ece2b5d62aa4e754f7f7c5e2cb722a29afe4fb88f1a10f341719e" gracePeriod=30 Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.244852 4815 scope.go:117] "RemoveContainer" containerID="cef0e7e70b9abdc983c0fa359e7d139f9e6011b9efe129abfd2d41f709ac71aa" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.253148 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44jd4\" (UniqueName: \"kubernetes.io/projected/ea8d6d97-73b0-4db9-8712-31eb46ef0105-kube-api-access-44jd4\") pod \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.253199 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-dns-swift-storage-0\") pod \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.253244 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-ovsdbserver-nb\") pod \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.253322 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-config\") pod \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.253468 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-dns-svc\") pod \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.253504 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-ovsdbserver-sb\") pod \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\" (UID: \"ea8d6d97-73b0-4db9-8712-31eb46ef0105\") " Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.269082 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea8d6d97-73b0-4db9-8712-31eb46ef0105-kube-api-access-44jd4" (OuterVolumeSpecName: "kube-api-access-44jd4") pod "ea8d6d97-73b0-4db9-8712-31eb46ef0105" (UID: "ea8d6d97-73b0-4db9-8712-31eb46ef0105"). InnerVolumeSpecName "kube-api-access-44jd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.324035 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.333360 4815 scope.go:117] "RemoveContainer" containerID="0c4fecacd29410071ab5edabaf794d418b4b3db1af96dc7c45b4fdc9554e3452" Feb 25 13:38:51 crc kubenswrapper[4815]: E0225 13:38:51.342076 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c4fecacd29410071ab5edabaf794d418b4b3db1af96dc7c45b4fdc9554e3452\": container with ID starting with 0c4fecacd29410071ab5edabaf794d418b4b3db1af96dc7c45b4fdc9554e3452 not found: ID does not exist" containerID="0c4fecacd29410071ab5edabaf794d418b4b3db1af96dc7c45b4fdc9554e3452" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.342295 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c4fecacd29410071ab5edabaf794d418b4b3db1af96dc7c45b4fdc9554e3452"} err="failed to get container status \"0c4fecacd29410071ab5edabaf794d418b4b3db1af96dc7c45b4fdc9554e3452\": rpc error: code = NotFound desc = could not find container \"0c4fecacd29410071ab5edabaf794d418b4b3db1af96dc7c45b4fdc9554e3452\": container with ID starting with 0c4fecacd29410071ab5edabaf794d418b4b3db1af96dc7c45b4fdc9554e3452 not found: ID does not exist" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.342376 4815 scope.go:117] "RemoveContainer" containerID="cef0e7e70b9abdc983c0fa359e7d139f9e6011b9efe129abfd2d41f709ac71aa" Feb 25 13:38:51 crc kubenswrapper[4815]: E0225 13:38:51.351755 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cef0e7e70b9abdc983c0fa359e7d139f9e6011b9efe129abfd2d41f709ac71aa\": container with ID starting with cef0e7e70b9abdc983c0fa359e7d139f9e6011b9efe129abfd2d41f709ac71aa not found: ID does not exist" containerID="cef0e7e70b9abdc983c0fa359e7d139f9e6011b9efe129abfd2d41f709ac71aa" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.351799 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cef0e7e70b9abdc983c0fa359e7d139f9e6011b9efe129abfd2d41f709ac71aa"} err="failed to get container status \"cef0e7e70b9abdc983c0fa359e7d139f9e6011b9efe129abfd2d41f709ac71aa\": rpc error: code = NotFound desc = could not find container \"cef0e7e70b9abdc983c0fa359e7d139f9e6011b9efe129abfd2d41f709ac71aa\": container with ID starting with cef0e7e70b9abdc983c0fa359e7d139f9e6011b9efe129abfd2d41f709ac71aa not found: ID does not exist" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.357673 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44jd4\" (UniqueName: \"kubernetes.io/projected/ea8d6d97-73b0-4db9-8712-31eb46ef0105-kube-api-access-44jd4\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.378699 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ea8d6d97-73b0-4db9-8712-31eb46ef0105" (UID: "ea8d6d97-73b0-4db9-8712-31eb46ef0105"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.408248 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ea8d6d97-73b0-4db9-8712-31eb46ef0105" (UID: "ea8d6d97-73b0-4db9-8712-31eb46ef0105"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.419074 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ea8d6d97-73b0-4db9-8712-31eb46ef0105" (UID: "ea8d6d97-73b0-4db9-8712-31eb46ef0105"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.440076 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ea8d6d97-73b0-4db9-8712-31eb46ef0105" (UID: "ea8d6d97-73b0-4db9-8712-31eb46ef0105"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.440621 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-config" (OuterVolumeSpecName: "config") pod "ea8d6d97-73b0-4db9-8712-31eb46ef0105" (UID: "ea8d6d97-73b0-4db9-8712-31eb46ef0105"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.459357 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.459398 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.459410 4815 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.459419 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.459427 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea8d6d97-73b0-4db9-8712-31eb46ef0105-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.553547 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-gggp2"] Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.560913 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-gggp2"] Feb 25 13:38:51 crc kubenswrapper[4815]: I0225 13:38:51.767112 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:38:52 crc kubenswrapper[4815]: I0225 13:38:52.230410 4815 generic.go:334] "Generic (PLEG): container finished" podID="9c7055a7-b147-4629-906c-ecfcc968733b" containerID="09af70168a6ece2b5d62aa4e754f7f7c5e2cb722a29afe4fb88f1a10f341719e" exitCode=0 Feb 25 13:38:52 crc kubenswrapper[4815]: I0225 13:38:52.230695 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9c7055a7-b147-4629-906c-ecfcc968733b","Type":"ContainerDied","Data":"09af70168a6ece2b5d62aa4e754f7f7c5e2cb722a29afe4fb88f1a10f341719e"} Feb 25 13:38:52 crc kubenswrapper[4815]: I0225 13:38:52.951146 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea8d6d97-73b0-4db9-8712-31eb46ef0105" path="/var/lib/kubelet/pods/ea8d6d97-73b0-4db9-8712-31eb46ef0105/volumes" Feb 25 13:38:53 crc kubenswrapper[4815]: I0225 13:38:53.243325 4815 generic.go:334] "Generic (PLEG): container finished" podID="5d076fbb-4499-456f-82ee-211a8909fc99" containerID="9fba6bb1e21ffc87dd876586c19b77148678fd408e930bd6088e1ed1c188d47d" exitCode=0 Feb 25 13:38:53 crc kubenswrapper[4815]: I0225 13:38:53.243369 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6497d456d-s5g59" event={"ID":"5d076fbb-4499-456f-82ee-211a8909fc99","Type":"ContainerDied","Data":"9fba6bb1e21ffc87dd876586c19b77148678fd408e930bd6088e1ed1c188d47d"} Feb 25 13:38:53 crc kubenswrapper[4815]: I0225 13:38:53.837831 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6497d456d-s5g59" podUID="5d076fbb-4499-456f-82ee-211a8909fc99" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 25 13:38:54 crc kubenswrapper[4815]: I0225 13:38:54.600335 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6cdd69d876-dmcfv" Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.795622 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.870317 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-config-data-custom\") pod \"9c7055a7-b147-4629-906c-ecfcc968733b\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.870380 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-combined-ca-bundle\") pod \"9c7055a7-b147-4629-906c-ecfcc968733b\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.870441 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c7055a7-b147-4629-906c-ecfcc968733b-etc-machine-id\") pod \"9c7055a7-b147-4629-906c-ecfcc968733b\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.870486 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-config-data\") pod \"9c7055a7-b147-4629-906c-ecfcc968733b\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.870603 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-scripts\") pod \"9c7055a7-b147-4629-906c-ecfcc968733b\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.870743 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlnjh\" (UniqueName: \"kubernetes.io/projected/9c7055a7-b147-4629-906c-ecfcc968733b-kube-api-access-jlnjh\") pod \"9c7055a7-b147-4629-906c-ecfcc968733b\" (UID: \"9c7055a7-b147-4629-906c-ecfcc968733b\") " Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.872365 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c7055a7-b147-4629-906c-ecfcc968733b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9c7055a7-b147-4629-906c-ecfcc968733b" (UID: "9c7055a7-b147-4629-906c-ecfcc968733b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.882670 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-scripts" (OuterVolumeSpecName: "scripts") pod "9c7055a7-b147-4629-906c-ecfcc968733b" (UID: "9c7055a7-b147-4629-906c-ecfcc968733b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.882769 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9c7055a7-b147-4629-906c-ecfcc968733b" (UID: "9c7055a7-b147-4629-906c-ecfcc968733b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.905744 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c7055a7-b147-4629-906c-ecfcc968733b-kube-api-access-jlnjh" (OuterVolumeSpecName: "kube-api-access-jlnjh") pod "9c7055a7-b147-4629-906c-ecfcc968733b" (UID: "9c7055a7-b147-4629-906c-ecfcc968733b"). InnerVolumeSpecName "kube-api-access-jlnjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.953397 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c7055a7-b147-4629-906c-ecfcc968733b" (UID: "9c7055a7-b147-4629-906c-ecfcc968733b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.974684 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.974720 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlnjh\" (UniqueName: \"kubernetes.io/projected/9c7055a7-b147-4629-906c-ecfcc968733b-kube-api-access-jlnjh\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.974736 4815 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.974747 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.974760 4815 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c7055a7-b147-4629-906c-ecfcc968733b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:55 crc kubenswrapper[4815]: I0225 13:38:55.995620 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-config-data" (OuterVolumeSpecName: "config-data") pod "9c7055a7-b147-4629-906c-ecfcc968733b" (UID: "9c7055a7-b147-4629-906c-ecfcc968733b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.076264 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c7055a7-b147-4629-906c-ecfcc968733b-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.272973 4815 generic.go:334] "Generic (PLEG): container finished" podID="9c7055a7-b147-4629-906c-ecfcc968733b" containerID="653acdd019d486b2d24004094ba6855b8e3416fd5cce26d5e7b10c7ae8129191" exitCode=0 Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.273015 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9c7055a7-b147-4629-906c-ecfcc968733b","Type":"ContainerDied","Data":"653acdd019d486b2d24004094ba6855b8e3416fd5cce26d5e7b10c7ae8129191"} Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.273046 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9c7055a7-b147-4629-906c-ecfcc968733b","Type":"ContainerDied","Data":"3f291da4e56c94bd06610767bfbc97a029f86de1d9a87fe1ebda5882d60f83e4"} Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.273067 4815 scope.go:117] "RemoveContainer" containerID="09af70168a6ece2b5d62aa4e754f7f7c5e2cb722a29afe4fb88f1a10f341719e" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.273077 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.320755 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.330716 4815 scope.go:117] "RemoveContainer" containerID="653acdd019d486b2d24004094ba6855b8e3416fd5cce26d5e7b10c7ae8129191" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.334051 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.346081 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 13:38:56 crc kubenswrapper[4815]: E0225 13:38:56.346427 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c7055a7-b147-4629-906c-ecfcc968733b" containerName="cinder-scheduler" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.346438 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c7055a7-b147-4629-906c-ecfcc968733b" containerName="cinder-scheduler" Feb 25 13:38:56 crc kubenswrapper[4815]: E0225 13:38:56.346453 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="145d5de3-c56d-4209-b3b7-348f2f6d2365" containerName="neutron-api" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.346459 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="145d5de3-c56d-4209-b3b7-348f2f6d2365" containerName="neutron-api" Feb 25 13:38:56 crc kubenswrapper[4815]: E0225 13:38:56.346471 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="145d5de3-c56d-4209-b3b7-348f2f6d2365" containerName="neutron-httpd" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.346477 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="145d5de3-c56d-4209-b3b7-348f2f6d2365" containerName="neutron-httpd" Feb 25 13:38:56 crc kubenswrapper[4815]: E0225 13:38:56.346494 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea8d6d97-73b0-4db9-8712-31eb46ef0105" containerName="init" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.346500 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea8d6d97-73b0-4db9-8712-31eb46ef0105" containerName="init" Feb 25 13:38:56 crc kubenswrapper[4815]: E0225 13:38:56.346526 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c7055a7-b147-4629-906c-ecfcc968733b" containerName="probe" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.346531 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c7055a7-b147-4629-906c-ecfcc968733b" containerName="probe" Feb 25 13:38:56 crc kubenswrapper[4815]: E0225 13:38:56.346551 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea8d6d97-73b0-4db9-8712-31eb46ef0105" containerName="dnsmasq-dns" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.346556 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea8d6d97-73b0-4db9-8712-31eb46ef0105" containerName="dnsmasq-dns" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.346718 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="145d5de3-c56d-4209-b3b7-348f2f6d2365" containerName="neutron-httpd" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.346728 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="145d5de3-c56d-4209-b3b7-348f2f6d2365" containerName="neutron-api" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.346737 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c7055a7-b147-4629-906c-ecfcc968733b" containerName="cinder-scheduler" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.346751 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c7055a7-b147-4629-906c-ecfcc968733b" containerName="probe" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.346762 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea8d6d97-73b0-4db9-8712-31eb46ef0105" containerName="dnsmasq-dns" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.356719 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.362839 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.379121 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.410769 4815 scope.go:117] "RemoveContainer" containerID="09af70168a6ece2b5d62aa4e754f7f7c5e2cb722a29afe4fb88f1a10f341719e" Feb 25 13:38:56 crc kubenswrapper[4815]: E0225 13:38:56.411945 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09af70168a6ece2b5d62aa4e754f7f7c5e2cb722a29afe4fb88f1a10f341719e\": container with ID starting with 09af70168a6ece2b5d62aa4e754f7f7c5e2cb722a29afe4fb88f1a10f341719e not found: ID does not exist" containerID="09af70168a6ece2b5d62aa4e754f7f7c5e2cb722a29afe4fb88f1a10f341719e" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.411980 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09af70168a6ece2b5d62aa4e754f7f7c5e2cb722a29afe4fb88f1a10f341719e"} err="failed to get container status \"09af70168a6ece2b5d62aa4e754f7f7c5e2cb722a29afe4fb88f1a10f341719e\": rpc error: code = NotFound desc = could not find container \"09af70168a6ece2b5d62aa4e754f7f7c5e2cb722a29afe4fb88f1a10f341719e\": container with ID starting with 09af70168a6ece2b5d62aa4e754f7f7c5e2cb722a29afe4fb88f1a10f341719e not found: ID does not exist" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.411999 4815 scope.go:117] "RemoveContainer" containerID="653acdd019d486b2d24004094ba6855b8e3416fd5cce26d5e7b10c7ae8129191" Feb 25 13:38:56 crc kubenswrapper[4815]: E0225 13:38:56.413000 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"653acdd019d486b2d24004094ba6855b8e3416fd5cce26d5e7b10c7ae8129191\": container with ID starting with 653acdd019d486b2d24004094ba6855b8e3416fd5cce26d5e7b10c7ae8129191 not found: ID does not exist" containerID="653acdd019d486b2d24004094ba6855b8e3416fd5cce26d5e7b10c7ae8129191" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.413024 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"653acdd019d486b2d24004094ba6855b8e3416fd5cce26d5e7b10c7ae8129191"} err="failed to get container status \"653acdd019d486b2d24004094ba6855b8e3416fd5cce26d5e7b10c7ae8129191\": rpc error: code = NotFound desc = could not find container \"653acdd019d486b2d24004094ba6855b8e3416fd5cce26d5e7b10c7ae8129191\": container with ID starting with 653acdd019d486b2d24004094ba6855b8e3416fd5cce26d5e7b10c7ae8129191 not found: ID does not exist" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.483501 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.483564 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.483888 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-config-data\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.484001 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q4w9\" (UniqueName: \"kubernetes.io/projected/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-kube-api-access-8q4w9\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.484080 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-scripts\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.484207 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.586277 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-config-data\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.586336 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q4w9\" (UniqueName: \"kubernetes.io/projected/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-kube-api-access-8q4w9\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.586371 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-scripts\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.586413 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.586463 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.586491 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.587217 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.592440 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-config-data\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.592978 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.599000 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-scripts\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.600020 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.621287 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q4w9\" (UniqueName: \"kubernetes.io/projected/64bb2d31-f562-4398-a0d6-c6bee9b9ba35-kube-api-access-8q4w9\") pod \"cinder-scheduler-0\" (UID: \"64bb2d31-f562-4398-a0d6-c6bee9b9ba35\") " pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.702532 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.774204 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.822982 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-b695645c6-vszf4" Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.898130 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-75dc9744b8-t8qpj"] Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.898365 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-75dc9744b8-t8qpj" podUID="afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" containerName="placement-log" containerID="cri-o://68c5a50e0da94b36c79a7d43ce763c4da7f48c6c14c0fcd8ec3edbb5a3a77956" gracePeriod=30 Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.898834 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-75dc9744b8-t8qpj" podUID="afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" containerName="placement-api" containerID="cri-o://2aac1dd5a8902bfcc26014020926a1705bb21bc2b6daa68d882f15c2e47a07a2" gracePeriod=30 Feb 25 13:38:56 crc kubenswrapper[4815]: I0225 13:38:56.950159 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c7055a7-b147-4629-906c-ecfcc968733b" path="/var/lib/kubelet/pods/9c7055a7-b147-4629-906c-ecfcc968733b/volumes" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.029693 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.030770 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.033191 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.033463 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-m7q4g" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.033707 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.072614 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.097783 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aab70915-6ce4-487b-ba58-2589dce33bbb-openstack-config-secret\") pod \"openstackclient\" (UID: \"aab70915-6ce4-487b-ba58-2589dce33bbb\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.097820 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv6bj\" (UniqueName: \"kubernetes.io/projected/aab70915-6ce4-487b-ba58-2589dce33bbb-kube-api-access-cv6bj\") pod \"openstackclient\" (UID: \"aab70915-6ce4-487b-ba58-2589dce33bbb\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.097874 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aab70915-6ce4-487b-ba58-2589dce33bbb-openstack-config\") pod \"openstackclient\" (UID: \"aab70915-6ce4-487b-ba58-2589dce33bbb\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.097905 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab70915-6ce4-487b-ba58-2589dce33bbb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"aab70915-6ce4-487b-ba58-2589dce33bbb\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.199731 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aab70915-6ce4-487b-ba58-2589dce33bbb-openstack-config-secret\") pod \"openstackclient\" (UID: \"aab70915-6ce4-487b-ba58-2589dce33bbb\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.199780 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv6bj\" (UniqueName: \"kubernetes.io/projected/aab70915-6ce4-487b-ba58-2589dce33bbb-kube-api-access-cv6bj\") pod \"openstackclient\" (UID: \"aab70915-6ce4-487b-ba58-2589dce33bbb\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.199840 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aab70915-6ce4-487b-ba58-2589dce33bbb-openstack-config\") pod \"openstackclient\" (UID: \"aab70915-6ce4-487b-ba58-2589dce33bbb\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.199874 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab70915-6ce4-487b-ba58-2589dce33bbb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"aab70915-6ce4-487b-ba58-2589dce33bbb\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.200477 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.204376 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aab70915-6ce4-487b-ba58-2589dce33bbb-openstack-config-secret\") pod \"openstackclient\" (UID: \"aab70915-6ce4-487b-ba58-2589dce33bbb\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.204998 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab70915-6ce4-487b-ba58-2589dce33bbb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"aab70915-6ce4-487b-ba58-2589dce33bbb\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.207549 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aab70915-6ce4-487b-ba58-2589dce33bbb-openstack-config\") pod \"openstackclient\" (UID: \"aab70915-6ce4-487b-ba58-2589dce33bbb\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.216846 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv6bj\" (UniqueName: \"kubernetes.io/projected/aab70915-6ce4-487b-ba58-2589dce33bbb-kube-api-access-cv6bj\") pod \"openstackclient\" (UID: \"aab70915-6ce4-487b-ba58-2589dce33bbb\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.293447 4815 generic.go:334] "Generic (PLEG): container finished" podID="afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" containerID="68c5a50e0da94b36c79a7d43ce763c4da7f48c6c14c0fcd8ec3edbb5a3a77956" exitCode=143 Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.293617 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75dc9744b8-t8qpj" event={"ID":"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4","Type":"ContainerDied","Data":"68c5a50e0da94b36c79a7d43ce763c4da7f48c6c14c0fcd8ec3edbb5a3a77956"} Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.301156 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"64bb2d31-f562-4398-a0d6-c6bee9b9ba35","Type":"ContainerStarted","Data":"7dabb4633e10a7dab1328eea8191335fe08839a506bdb50864bc00c7c800c731"} Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.376812 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.377468 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.394181 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.411076 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.413049 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.431100 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.505682 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3f935809-d8cf-4a5f-8d6b-05d7045b6bc8-openstack-config-secret\") pod \"openstackclient\" (UID: \"3f935809-d8cf-4a5f-8d6b-05d7045b6bc8\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.505780 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds78g\" (UniqueName: \"kubernetes.io/projected/3f935809-d8cf-4a5f-8d6b-05d7045b6bc8-kube-api-access-ds78g\") pod \"openstackclient\" (UID: \"3f935809-d8cf-4a5f-8d6b-05d7045b6bc8\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.505835 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3f935809-d8cf-4a5f-8d6b-05d7045b6bc8-openstack-config\") pod \"openstackclient\" (UID: \"3f935809-d8cf-4a5f-8d6b-05d7045b6bc8\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.505976 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f935809-d8cf-4a5f-8d6b-05d7045b6bc8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3f935809-d8cf-4a5f-8d6b-05d7045b6bc8\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.542069 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:57 crc kubenswrapper[4815]: E0225 13:38:57.554756 4815 log.go:32] "RunPodSandbox from runtime service failed" err=< Feb 25 13:38:57 crc kubenswrapper[4815]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_aab70915-6ce4-487b-ba58-2589dce33bbb_0(1aba1f4a54ab042e1c43dbe6574e12cc2bf0b8cfd42b787410aa81329153e185): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"1aba1f4a54ab042e1c43dbe6574e12cc2bf0b8cfd42b787410aa81329153e185" Netns:"/var/run/netns/f5f065e1-f714-4b55-b9a5-46f65bc521f8" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=1aba1f4a54ab042e1c43dbe6574e12cc2bf0b8cfd42b787410aa81329153e185;K8S_POD_UID=aab70915-6ce4-487b-ba58-2589dce33bbb" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/aab70915-6ce4-487b-ba58-2589dce33bbb]: expected pod UID "aab70915-6ce4-487b-ba58-2589dce33bbb" but got "3f935809-d8cf-4a5f-8d6b-05d7045b6bc8" from Kube API Feb 25 13:38:57 crc kubenswrapper[4815]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 25 13:38:57 crc kubenswrapper[4815]: > Feb 25 13:38:57 crc kubenswrapper[4815]: E0225 13:38:57.554838 4815 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Feb 25 13:38:57 crc kubenswrapper[4815]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_aab70915-6ce4-487b-ba58-2589dce33bbb_0(1aba1f4a54ab042e1c43dbe6574e12cc2bf0b8cfd42b787410aa81329153e185): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"1aba1f4a54ab042e1c43dbe6574e12cc2bf0b8cfd42b787410aa81329153e185" Netns:"/var/run/netns/f5f065e1-f714-4b55-b9a5-46f65bc521f8" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=1aba1f4a54ab042e1c43dbe6574e12cc2bf0b8cfd42b787410aa81329153e185;K8S_POD_UID=aab70915-6ce4-487b-ba58-2589dce33bbb" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/aab70915-6ce4-487b-ba58-2589dce33bbb]: expected pod UID "aab70915-6ce4-487b-ba58-2589dce33bbb" but got "3f935809-d8cf-4a5f-8d6b-05d7045b6bc8" from Kube API Feb 25 13:38:57 crc kubenswrapper[4815]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 25 13:38:57 crc kubenswrapper[4815]: > pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.609542 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3f935809-d8cf-4a5f-8d6b-05d7045b6bc8-openstack-config-secret\") pod \"openstackclient\" (UID: \"3f935809-d8cf-4a5f-8d6b-05d7045b6bc8\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.609635 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds78g\" (UniqueName: \"kubernetes.io/projected/3f935809-d8cf-4a5f-8d6b-05d7045b6bc8-kube-api-access-ds78g\") pod \"openstackclient\" (UID: \"3f935809-d8cf-4a5f-8d6b-05d7045b6bc8\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.609671 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3f935809-d8cf-4a5f-8d6b-05d7045b6bc8-openstack-config\") pod \"openstackclient\" (UID: \"3f935809-d8cf-4a5f-8d6b-05d7045b6bc8\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.609756 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f935809-d8cf-4a5f-8d6b-05d7045b6bc8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3f935809-d8cf-4a5f-8d6b-05d7045b6bc8\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.611301 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3f935809-d8cf-4a5f-8d6b-05d7045b6bc8-openstack-config\") pod \"openstackclient\" (UID: \"3f935809-d8cf-4a5f-8d6b-05d7045b6bc8\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.614716 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f935809-d8cf-4a5f-8d6b-05d7045b6bc8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3f935809-d8cf-4a5f-8d6b-05d7045b6bc8\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.616920 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3f935809-d8cf-4a5f-8d6b-05d7045b6bc8-openstack-config-secret\") pod \"openstackclient\" (UID: \"3f935809-d8cf-4a5f-8d6b-05d7045b6bc8\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.629954 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds78g\" (UniqueName: \"kubernetes.io/projected/3f935809-d8cf-4a5f-8d6b-05d7045b6bc8-kube-api-access-ds78g\") pod \"openstackclient\" (UID: \"3f935809-d8cf-4a5f-8d6b-05d7045b6bc8\") " pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.795858 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 13:38:57 crc kubenswrapper[4815]: I0225 13:38:57.962003 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7fc6fdc978-tzrxw" Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.050846 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.063517 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-b86d988f8-9zdkl"] Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.063746 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-b86d988f8-9zdkl" podUID="8114370d-9aee-4979-9445-2f0f57e805cf" containerName="barbican-api-log" containerID="cri-o://54356262d93d48a51a0e4589f1246a5b05e27e0966783a0604b95de28ec29155" gracePeriod=30 Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.063872 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-b86d988f8-9zdkl" podUID="8114370d-9aee-4979-9445-2f0f57e805cf" containerName="barbican-api" containerID="cri-o://7989163bcdd6a0940634a28e03ec0e2678fb563e455feefb7f59ab23861a70f3" gracePeriod=30 Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.317690 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.321525 4815 generic.go:334] "Generic (PLEG): container finished" podID="8114370d-9aee-4979-9445-2f0f57e805cf" containerID="54356262d93d48a51a0e4589f1246a5b05e27e0966783a0604b95de28ec29155" exitCode=143 Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.321644 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b86d988f8-9zdkl" event={"ID":"8114370d-9aee-4979-9445-2f0f57e805cf","Type":"ContainerDied","Data":"54356262d93d48a51a0e4589f1246a5b05e27e0966783a0604b95de28ec29155"} Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.323258 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.323878 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"64bb2d31-f562-4398-a0d6-c6bee9b9ba35","Type":"ContainerStarted","Data":"b75cf0341ad5059c61febb06324967d556aa3d59b33f488f246df2ca8f5a86b2"} Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.330857 4815 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="aab70915-6ce4-487b-ba58-2589dce33bbb" podUID="3f935809-d8cf-4a5f-8d6b-05d7045b6bc8" Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.396703 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.529899 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab70915-6ce4-487b-ba58-2589dce33bbb-combined-ca-bundle\") pod \"aab70915-6ce4-487b-ba58-2589dce33bbb\" (UID: \"aab70915-6ce4-487b-ba58-2589dce33bbb\") " Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.530426 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aab70915-6ce4-487b-ba58-2589dce33bbb-openstack-config\") pod \"aab70915-6ce4-487b-ba58-2589dce33bbb\" (UID: \"aab70915-6ce4-487b-ba58-2589dce33bbb\") " Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.530728 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aab70915-6ce4-487b-ba58-2589dce33bbb-openstack-config-secret\") pod \"aab70915-6ce4-487b-ba58-2589dce33bbb\" (UID: \"aab70915-6ce4-487b-ba58-2589dce33bbb\") " Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.530797 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cv6bj\" (UniqueName: \"kubernetes.io/projected/aab70915-6ce4-487b-ba58-2589dce33bbb-kube-api-access-cv6bj\") pod \"aab70915-6ce4-487b-ba58-2589dce33bbb\" (UID: \"aab70915-6ce4-487b-ba58-2589dce33bbb\") " Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.531417 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aab70915-6ce4-487b-ba58-2589dce33bbb-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "aab70915-6ce4-487b-ba58-2589dce33bbb" (UID: "aab70915-6ce4-487b-ba58-2589dce33bbb"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.531630 4815 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aab70915-6ce4-487b-ba58-2589dce33bbb-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.535252 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab70915-6ce4-487b-ba58-2589dce33bbb-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "aab70915-6ce4-487b-ba58-2589dce33bbb" (UID: "aab70915-6ce4-487b-ba58-2589dce33bbb"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.540151 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aab70915-6ce4-487b-ba58-2589dce33bbb-kube-api-access-cv6bj" (OuterVolumeSpecName: "kube-api-access-cv6bj") pod "aab70915-6ce4-487b-ba58-2589dce33bbb" (UID: "aab70915-6ce4-487b-ba58-2589dce33bbb"). InnerVolumeSpecName "kube-api-access-cv6bj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.555648 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab70915-6ce4-487b-ba58-2589dce33bbb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aab70915-6ce4-487b-ba58-2589dce33bbb" (UID: "aab70915-6ce4-487b-ba58-2589dce33bbb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.634346 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab70915-6ce4-487b-ba58-2589dce33bbb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.634376 4815 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aab70915-6ce4-487b-ba58-2589dce33bbb-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.634388 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cv6bj\" (UniqueName: \"kubernetes.io/projected/aab70915-6ce4-487b-ba58-2589dce33bbb-kube-api-access-cv6bj\") on node \"crc\" DevicePath \"\"" Feb 25 13:38:58 crc kubenswrapper[4815]: I0225 13:38:58.945585 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aab70915-6ce4-487b-ba58-2589dce33bbb" path="/var/lib/kubelet/pods/aab70915-6ce4-487b-ba58-2589dce33bbb/volumes" Feb 25 13:38:59 crc kubenswrapper[4815]: I0225 13:38:59.334600 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"3f935809-d8cf-4a5f-8d6b-05d7045b6bc8","Type":"ContainerStarted","Data":"e91acc9f38a92eebf9db54dd925041147676474a5aeb3eaa611eb7197ed0cf9e"} Feb 25 13:38:59 crc kubenswrapper[4815]: I0225 13:38:59.337027 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 13:38:59 crc kubenswrapper[4815]: I0225 13:38:59.337159 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"64bb2d31-f562-4398-a0d6-c6bee9b9ba35","Type":"ContainerStarted","Data":"9e9645321e38c7eb15ef48fdfc8d03df13401bff2e4da51ca92a3675ed503d89"} Feb 25 13:38:59 crc kubenswrapper[4815]: I0225 13:38:59.365914 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.36589845 podStartE2EDuration="3.36589845s" podCreationTimestamp="2026-02-25 13:38:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:38:59.363266043 +0000 UTC m=+1097.164364117" watchObservedRunningTime="2026-02-25 13:38:59.36589845 +0000 UTC m=+1097.166996504" Feb 25 13:38:59 crc kubenswrapper[4815]: I0225 13:38:59.365983 4815 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="aab70915-6ce4-487b-ba58-2589dce33bbb" podUID="3f935809-d8cf-4a5f-8d6b-05d7045b6bc8" Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.382283 4815 generic.go:334] "Generic (PLEG): container finished" podID="afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" containerID="2aac1dd5a8902bfcc26014020926a1705bb21bc2b6daa68d882f15c2e47a07a2" exitCode=0 Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.383303 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75dc9744b8-t8qpj" event={"ID":"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4","Type":"ContainerDied","Data":"2aac1dd5a8902bfcc26014020926a1705bb21bc2b6daa68d882f15c2e47a07a2"} Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.567616 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.668174 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-scripts\") pod \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.668212 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-public-tls-certs\") pod \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.668239 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-internal-tls-certs\") pod \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.668260 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-combined-ca-bundle\") pod \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.668292 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-config-data\") pod \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.668399 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-logs\") pod \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.668499 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nv4p\" (UniqueName: \"kubernetes.io/projected/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-kube-api-access-6nv4p\") pod \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\" (UID: \"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4\") " Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.669600 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-logs" (OuterVolumeSpecName: "logs") pod "afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" (UID: "afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.690235 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-scripts" (OuterVolumeSpecName: "scripts") pod "afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" (UID: "afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.698693 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-kube-api-access-6nv4p" (OuterVolumeSpecName: "kube-api-access-6nv4p") pod "afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" (UID: "afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4"). InnerVolumeSpecName "kube-api-access-6nv4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.722727 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-config-data" (OuterVolumeSpecName: "config-data") pod "afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" (UID: "afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.734140 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" (UID: "afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.770723 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nv4p\" (UniqueName: \"kubernetes.io/projected/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-kube-api-access-6nv4p\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.770759 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.770769 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.770777 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.770788 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.787816 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" (UID: "afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.789888 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" (UID: "afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.872082 4815 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:00 crc kubenswrapper[4815]: I0225 13:39:00.872121 4815 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:01 crc kubenswrapper[4815]: E0225 13:39:01.111712 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafdec2fd_f5ed_49ce_8ce0_4e5b0e2b7fe4.slice/crio-ee6a4dcb6b577fb8a0d21147dce77741bb4cad2771cab8604d803f2efc4271f9\": RecentStats: unable to find data in memory cache]" Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.244268 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-b86d988f8-9zdkl" podUID="8114370d-9aee-4979-9445-2f0f57e805cf" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.168:9311/healthcheck\": read tcp 10.217.0.2:34668->10.217.0.168:9311: read: connection reset by peer" Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.245005 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-b86d988f8-9zdkl" podUID="8114370d-9aee-4979-9445-2f0f57e805cf" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.168:9311/healthcheck\": read tcp 10.217.0.2:34654->10.217.0.168:9311: read: connection reset by peer" Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.407599 4815 generic.go:334] "Generic (PLEG): container finished" podID="8114370d-9aee-4979-9445-2f0f57e805cf" containerID="7989163bcdd6a0940634a28e03ec0e2678fb563e455feefb7f59ab23861a70f3" exitCode=0 Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.407957 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b86d988f8-9zdkl" event={"ID":"8114370d-9aee-4979-9445-2f0f57e805cf","Type":"ContainerDied","Data":"7989163bcdd6a0940634a28e03ec0e2678fb563e455feefb7f59ab23861a70f3"} Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.410417 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75dc9744b8-t8qpj" event={"ID":"afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4","Type":"ContainerDied","Data":"ee6a4dcb6b577fb8a0d21147dce77741bb4cad2771cab8604d803f2efc4271f9"} Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.410469 4815 scope.go:117] "RemoveContainer" containerID="2aac1dd5a8902bfcc26014020926a1705bb21bc2b6daa68d882f15c2e47a07a2" Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.410652 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75dc9744b8-t8qpj" Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.443416 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-75dc9744b8-t8qpj"] Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.451895 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-75dc9744b8-t8qpj"] Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.456260 4815 scope.go:117] "RemoveContainer" containerID="68c5a50e0da94b36c79a7d43ce763c4da7f48c6c14c0fcd8ec3edbb5a3a77956" Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.703408 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.704134 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.787126 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-combined-ca-bundle\") pod \"8114370d-9aee-4979-9445-2f0f57e805cf\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.787257 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7sdl\" (UniqueName: \"kubernetes.io/projected/8114370d-9aee-4979-9445-2f0f57e805cf-kube-api-access-t7sdl\") pod \"8114370d-9aee-4979-9445-2f0f57e805cf\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.787285 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-config-data-custom\") pod \"8114370d-9aee-4979-9445-2f0f57e805cf\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.787339 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-config-data\") pod \"8114370d-9aee-4979-9445-2f0f57e805cf\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.787370 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8114370d-9aee-4979-9445-2f0f57e805cf-logs\") pod \"8114370d-9aee-4979-9445-2f0f57e805cf\" (UID: \"8114370d-9aee-4979-9445-2f0f57e805cf\") " Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.788750 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8114370d-9aee-4979-9445-2f0f57e805cf-logs" (OuterVolumeSpecName: "logs") pod "8114370d-9aee-4979-9445-2f0f57e805cf" (UID: "8114370d-9aee-4979-9445-2f0f57e805cf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.792742 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8114370d-9aee-4979-9445-2f0f57e805cf-kube-api-access-t7sdl" (OuterVolumeSpecName: "kube-api-access-t7sdl") pod "8114370d-9aee-4979-9445-2f0f57e805cf" (UID: "8114370d-9aee-4979-9445-2f0f57e805cf"). InnerVolumeSpecName "kube-api-access-t7sdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.821298 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8114370d-9aee-4979-9445-2f0f57e805cf" (UID: "8114370d-9aee-4979-9445-2f0f57e805cf"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.838891 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8114370d-9aee-4979-9445-2f0f57e805cf" (UID: "8114370d-9aee-4979-9445-2f0f57e805cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.865530 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-config-data" (OuterVolumeSpecName: "config-data") pod "8114370d-9aee-4979-9445-2f0f57e805cf" (UID: "8114370d-9aee-4979-9445-2f0f57e805cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.889642 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.889680 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7sdl\" (UniqueName: \"kubernetes.io/projected/8114370d-9aee-4979-9445-2f0f57e805cf-kube-api-access-t7sdl\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.889692 4815 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.889702 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8114370d-9aee-4979-9445-2f0f57e805cf-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:01 crc kubenswrapper[4815]: I0225 13:39:01.889714 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8114370d-9aee-4979-9445-2f0f57e805cf-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.419115 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b86d988f8-9zdkl" event={"ID":"8114370d-9aee-4979-9445-2f0f57e805cf","Type":"ContainerDied","Data":"7d9efe2a61521b7fe2b1800cd02c7d39f0b57a0806ac39e11741992588bad282"} Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.419166 4815 scope.go:117] "RemoveContainer" containerID="7989163bcdd6a0940634a28e03ec0e2678fb563e455feefb7f59ab23861a70f3" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.419257 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-b86d988f8-9zdkl" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.454368 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-b86d988f8-9zdkl"] Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.462170 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-b86d988f8-9zdkl"] Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.468854 4815 scope.go:117] "RemoveContainer" containerID="54356262d93d48a51a0e4589f1246a5b05e27e0966783a0604b95de28ec29155" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.528723 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-68d8cd5787-kmnfm"] Feb 25 13:39:02 crc kubenswrapper[4815]: E0225 13:39:02.529260 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8114370d-9aee-4979-9445-2f0f57e805cf" containerName="barbican-api" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.529276 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8114370d-9aee-4979-9445-2f0f57e805cf" containerName="barbican-api" Feb 25 13:39:02 crc kubenswrapper[4815]: E0225 13:39:02.529287 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8114370d-9aee-4979-9445-2f0f57e805cf" containerName="barbican-api-log" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.529297 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8114370d-9aee-4979-9445-2f0f57e805cf" containerName="barbican-api-log" Feb 25 13:39:02 crc kubenswrapper[4815]: E0225 13:39:02.529309 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" containerName="placement-api" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.529315 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" containerName="placement-api" Feb 25 13:39:02 crc kubenswrapper[4815]: E0225 13:39:02.529335 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" containerName="placement-log" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.529340 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" containerName="placement-log" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.529489 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8114370d-9aee-4979-9445-2f0f57e805cf" containerName="barbican-api-log" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.529542 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" containerName="placement-api" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.529562 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" containerName="placement-log" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.529572 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8114370d-9aee-4979-9445-2f0f57e805cf" containerName="barbican-api" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.552719 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-68d8cd5787-kmnfm"] Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.552834 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.561370 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.561392 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.562185 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.603114 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-444h5\" (UniqueName: \"kubernetes.io/projected/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-kube-api-access-444h5\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.603187 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-config-data\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.603220 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-internal-tls-certs\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.603243 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-combined-ca-bundle\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.603282 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-run-httpd\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.603305 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-etc-swift\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.603333 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-public-tls-certs\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.603359 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-log-httpd\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.704877 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-public-tls-certs\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.704944 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-log-httpd\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.704984 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-444h5\" (UniqueName: \"kubernetes.io/projected/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-kube-api-access-444h5\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.705031 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-config-data\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.705065 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-internal-tls-certs\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.705089 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-combined-ca-bundle\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.705127 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-run-httpd\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.705152 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-etc-swift\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.705674 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-log-httpd\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.706320 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-run-httpd\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.710388 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-combined-ca-bundle\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.711117 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-internal-tls-certs\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.711392 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-config-data\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.711422 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-public-tls-certs\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.712427 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-etc-swift\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.721429 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-444h5\" (UniqueName: \"kubernetes.io/projected/ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96-kube-api-access-444h5\") pod \"swift-proxy-68d8cd5787-kmnfm\" (UID: \"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96\") " pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.869988 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.947018 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8114370d-9aee-4979-9445-2f0f57e805cf" path="/var/lib/kubelet/pods/8114370d-9aee-4979-9445-2f0f57e805cf/volumes" Feb 25 13:39:02 crc kubenswrapper[4815]: I0225 13:39:02.947858 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4" path="/var/lib/kubelet/pods/afdec2fd-f5ed-49ce-8ce0-4e5b0e2b7fe4/volumes" Feb 25 13:39:03 crc kubenswrapper[4815]: I0225 13:39:03.837458 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6497d456d-s5g59" podUID="5d076fbb-4499-456f-82ee-211a8909fc99" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 25 13:39:03 crc kubenswrapper[4815]: I0225 13:39:03.973758 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:03 crc kubenswrapper[4815]: I0225 13:39:03.974074 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerName="ceilometer-central-agent" containerID="cri-o://dfe32d3f63f2e76c406dc47bcfa57eb2c27bc3a91d00b5753fa4909f18a475b3" gracePeriod=30 Feb 25 13:39:03 crc kubenswrapper[4815]: I0225 13:39:03.974883 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerName="proxy-httpd" containerID="cri-o://64fa5eafb58ae0796dbda8aa80d9b18239391d9bff07332f0ee6268798453139" gracePeriod=30 Feb 25 13:39:03 crc kubenswrapper[4815]: I0225 13:39:03.974915 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerName="ceilometer-notification-agent" containerID="cri-o://0e3f2c6c27b9e6a889af97e0163bfa7460186ec10b5b75dc96cf6c63752c31b3" gracePeriod=30 Feb 25 13:39:03 crc kubenswrapper[4815]: I0225 13:39:03.974896 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerName="sg-core" containerID="cri-o://d7e37413d82a64dba8f1ff0c987f3cc5793ccc301f538991657226e2a5025542" gracePeriod=30 Feb 25 13:39:03 crc kubenswrapper[4815]: I0225 13:39:03.982333 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 25 13:39:04 crc kubenswrapper[4815]: I0225 13:39:04.462190 4815 generic.go:334] "Generic (PLEG): container finished" podID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerID="64fa5eafb58ae0796dbda8aa80d9b18239391d9bff07332f0ee6268798453139" exitCode=0 Feb 25 13:39:04 crc kubenswrapper[4815]: I0225 13:39:04.462228 4815 generic.go:334] "Generic (PLEG): container finished" podID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerID="d7e37413d82a64dba8f1ff0c987f3cc5793ccc301f538991657226e2a5025542" exitCode=2 Feb 25 13:39:04 crc kubenswrapper[4815]: I0225 13:39:04.462238 4815 generic.go:334] "Generic (PLEG): container finished" podID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerID="dfe32d3f63f2e76c406dc47bcfa57eb2c27bc3a91d00b5753fa4909f18a475b3" exitCode=0 Feb 25 13:39:04 crc kubenswrapper[4815]: I0225 13:39:04.462260 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"178df77c-9b85-4f6e-8c47-44f47dc60383","Type":"ContainerDied","Data":"64fa5eafb58ae0796dbda8aa80d9b18239391d9bff07332f0ee6268798453139"} Feb 25 13:39:04 crc kubenswrapper[4815]: I0225 13:39:04.462292 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"178df77c-9b85-4f6e-8c47-44f47dc60383","Type":"ContainerDied","Data":"d7e37413d82a64dba8f1ff0c987f3cc5793ccc301f538991657226e2a5025542"} Feb 25 13:39:04 crc kubenswrapper[4815]: I0225 13:39:04.462306 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"178df77c-9b85-4f6e-8c47-44f47dc60383","Type":"ContainerDied","Data":"dfe32d3f63f2e76c406dc47bcfa57eb2c27bc3a91d00b5753fa4909f18a475b3"} Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.521084 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-gcl8c"] Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.522893 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-gcl8c" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.536901 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-gcl8c"] Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.553996 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjt46\" (UniqueName: \"kubernetes.io/projected/2da1710b-375f-4795-993d-730ccbd04ccc-kube-api-access-kjt46\") pod \"nova-api-db-create-gcl8c\" (UID: \"2da1710b-375f-4795-993d-730ccbd04ccc\") " pod="openstack/nova-api-db-create-gcl8c" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.554063 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2da1710b-375f-4795-993d-730ccbd04ccc-operator-scripts\") pod \"nova-api-db-create-gcl8c\" (UID: \"2da1710b-375f-4795-993d-730ccbd04ccc\") " pod="openstack/nova-api-db-create-gcl8c" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.617852 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-p95dh"] Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.625143 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-p95dh" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.646175 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-p95dh"] Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.657714 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjjz4\" (UniqueName: \"kubernetes.io/projected/486b281d-a6fc-4c69-860a-ad0d5480fc77-kube-api-access-vjjz4\") pod \"nova-cell0-db-create-p95dh\" (UID: \"486b281d-a6fc-4c69-860a-ad0d5480fc77\") " pod="openstack/nova-cell0-db-create-p95dh" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.658022 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/486b281d-a6fc-4c69-860a-ad0d5480fc77-operator-scripts\") pod \"nova-cell0-db-create-p95dh\" (UID: \"486b281d-a6fc-4c69-860a-ad0d5480fc77\") " pod="openstack/nova-cell0-db-create-p95dh" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.658120 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjt46\" (UniqueName: \"kubernetes.io/projected/2da1710b-375f-4795-993d-730ccbd04ccc-kube-api-access-kjt46\") pod \"nova-api-db-create-gcl8c\" (UID: \"2da1710b-375f-4795-993d-730ccbd04ccc\") " pod="openstack/nova-api-db-create-gcl8c" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.658231 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2da1710b-375f-4795-993d-730ccbd04ccc-operator-scripts\") pod \"nova-api-db-create-gcl8c\" (UID: \"2da1710b-375f-4795-993d-730ccbd04ccc\") " pod="openstack/nova-api-db-create-gcl8c" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.659208 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2da1710b-375f-4795-993d-730ccbd04ccc-operator-scripts\") pod \"nova-api-db-create-gcl8c\" (UID: \"2da1710b-375f-4795-993d-730ccbd04ccc\") " pod="openstack/nova-api-db-create-gcl8c" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.675467 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjt46\" (UniqueName: \"kubernetes.io/projected/2da1710b-375f-4795-993d-730ccbd04ccc-kube-api-access-kjt46\") pod \"nova-api-db-create-gcl8c\" (UID: \"2da1710b-375f-4795-993d-730ccbd04ccc\") " pod="openstack/nova-api-db-create-gcl8c" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.733887 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0d2e-account-create-update-c7zn7"] Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.735004 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0d2e-account-create-update-c7zn7" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.737935 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.743383 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0d2e-account-create-update-c7zn7"] Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.761501 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/486b281d-a6fc-4c69-860a-ad0d5480fc77-operator-scripts\") pod \"nova-cell0-db-create-p95dh\" (UID: \"486b281d-a6fc-4c69-860a-ad0d5480fc77\") " pod="openstack/nova-cell0-db-create-p95dh" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.761657 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjjz4\" (UniqueName: \"kubernetes.io/projected/486b281d-a6fc-4c69-860a-ad0d5480fc77-kube-api-access-vjjz4\") pod \"nova-cell0-db-create-p95dh\" (UID: \"486b281d-a6fc-4c69-860a-ad0d5480fc77\") " pod="openstack/nova-cell0-db-create-p95dh" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.762974 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/486b281d-a6fc-4c69-860a-ad0d5480fc77-operator-scripts\") pod \"nova-cell0-db-create-p95dh\" (UID: \"486b281d-a6fc-4c69-860a-ad0d5480fc77\") " pod="openstack/nova-cell0-db-create-p95dh" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.778584 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjjz4\" (UniqueName: \"kubernetes.io/projected/486b281d-a6fc-4c69-860a-ad0d5480fc77-kube-api-access-vjjz4\") pod \"nova-cell0-db-create-p95dh\" (UID: \"486b281d-a6fc-4c69-860a-ad0d5480fc77\") " pod="openstack/nova-cell0-db-create-p95dh" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.832930 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-jldcn"] Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.834244 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jldcn" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.848528 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-jldcn"] Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.853388 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-gcl8c" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.863330 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dkpp\" (UniqueName: \"kubernetes.io/projected/7d139922-3032-4e41-978e-6bdd8ad8e0e1-kube-api-access-9dkpp\") pod \"nova-api-0d2e-account-create-update-c7zn7\" (UID: \"7d139922-3032-4e41-978e-6bdd8ad8e0e1\") " pod="openstack/nova-api-0d2e-account-create-update-c7zn7" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.863393 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d139922-3032-4e41-978e-6bdd8ad8e0e1-operator-scripts\") pod \"nova-api-0d2e-account-create-update-c7zn7\" (UID: \"7d139922-3032-4e41-978e-6bdd8ad8e0e1\") " pod="openstack/nova-api-0d2e-account-create-update-c7zn7" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.928650 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-5784-account-create-update-4gwt4"] Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.929980 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5784-account-create-update-4gwt4" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.934288 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.944966 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-p95dh" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.951354 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-5784-account-create-update-4gwt4"] Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.964819 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdnbv\" (UniqueName: \"kubernetes.io/projected/6648a53c-a316-42aa-8799-8a1decdb704c-kube-api-access-jdnbv\") pod \"nova-cell0-5784-account-create-update-4gwt4\" (UID: \"6648a53c-a316-42aa-8799-8a1decdb704c\") " pod="openstack/nova-cell0-5784-account-create-update-4gwt4" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.965093 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/036fbace-3dcf-484d-958d-10970b9c7177-operator-scripts\") pod \"nova-cell1-db-create-jldcn\" (UID: \"036fbace-3dcf-484d-958d-10970b9c7177\") " pod="openstack/nova-cell1-db-create-jldcn" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.965214 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6648a53c-a316-42aa-8799-8a1decdb704c-operator-scripts\") pod \"nova-cell0-5784-account-create-update-4gwt4\" (UID: \"6648a53c-a316-42aa-8799-8a1decdb704c\") " pod="openstack/nova-cell0-5784-account-create-update-4gwt4" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.965325 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ngsz\" (UniqueName: \"kubernetes.io/projected/036fbace-3dcf-484d-958d-10970b9c7177-kube-api-access-7ngsz\") pod \"nova-cell1-db-create-jldcn\" (UID: \"036fbace-3dcf-484d-958d-10970b9c7177\") " pod="openstack/nova-cell1-db-create-jldcn" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.965453 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dkpp\" (UniqueName: \"kubernetes.io/projected/7d139922-3032-4e41-978e-6bdd8ad8e0e1-kube-api-access-9dkpp\") pod \"nova-api-0d2e-account-create-update-c7zn7\" (UID: \"7d139922-3032-4e41-978e-6bdd8ad8e0e1\") " pod="openstack/nova-api-0d2e-account-create-update-c7zn7" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.965584 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d139922-3032-4e41-978e-6bdd8ad8e0e1-operator-scripts\") pod \"nova-api-0d2e-account-create-update-c7zn7\" (UID: \"7d139922-3032-4e41-978e-6bdd8ad8e0e1\") " pod="openstack/nova-api-0d2e-account-create-update-c7zn7" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.966213 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d139922-3032-4e41-978e-6bdd8ad8e0e1-operator-scripts\") pod \"nova-api-0d2e-account-create-update-c7zn7\" (UID: \"7d139922-3032-4e41-978e-6bdd8ad8e0e1\") " pod="openstack/nova-api-0d2e-account-create-update-c7zn7" Feb 25 13:39:05 crc kubenswrapper[4815]: I0225 13:39:05.983895 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dkpp\" (UniqueName: \"kubernetes.io/projected/7d139922-3032-4e41-978e-6bdd8ad8e0e1-kube-api-access-9dkpp\") pod \"nova-api-0d2e-account-create-update-c7zn7\" (UID: \"7d139922-3032-4e41-978e-6bdd8ad8e0e1\") " pod="openstack/nova-api-0d2e-account-create-update-c7zn7" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.054030 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0d2e-account-create-update-c7zn7" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.067702 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdnbv\" (UniqueName: \"kubernetes.io/projected/6648a53c-a316-42aa-8799-8a1decdb704c-kube-api-access-jdnbv\") pod \"nova-cell0-5784-account-create-update-4gwt4\" (UID: \"6648a53c-a316-42aa-8799-8a1decdb704c\") " pod="openstack/nova-cell0-5784-account-create-update-4gwt4" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.067754 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/036fbace-3dcf-484d-958d-10970b9c7177-operator-scripts\") pod \"nova-cell1-db-create-jldcn\" (UID: \"036fbace-3dcf-484d-958d-10970b9c7177\") " pod="openstack/nova-cell1-db-create-jldcn" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.067800 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6648a53c-a316-42aa-8799-8a1decdb704c-operator-scripts\") pod \"nova-cell0-5784-account-create-update-4gwt4\" (UID: \"6648a53c-a316-42aa-8799-8a1decdb704c\") " pod="openstack/nova-cell0-5784-account-create-update-4gwt4" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.067850 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ngsz\" (UniqueName: \"kubernetes.io/projected/036fbace-3dcf-484d-958d-10970b9c7177-kube-api-access-7ngsz\") pod \"nova-cell1-db-create-jldcn\" (UID: \"036fbace-3dcf-484d-958d-10970b9c7177\") " pod="openstack/nova-cell1-db-create-jldcn" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.068876 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6648a53c-a316-42aa-8799-8a1decdb704c-operator-scripts\") pod \"nova-cell0-5784-account-create-update-4gwt4\" (UID: \"6648a53c-a316-42aa-8799-8a1decdb704c\") " pod="openstack/nova-cell0-5784-account-create-update-4gwt4" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.077007 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/036fbace-3dcf-484d-958d-10970b9c7177-operator-scripts\") pod \"nova-cell1-db-create-jldcn\" (UID: \"036fbace-3dcf-484d-958d-10970b9c7177\") " pod="openstack/nova-cell1-db-create-jldcn" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.085264 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdnbv\" (UniqueName: \"kubernetes.io/projected/6648a53c-a316-42aa-8799-8a1decdb704c-kube-api-access-jdnbv\") pod \"nova-cell0-5784-account-create-update-4gwt4\" (UID: \"6648a53c-a316-42aa-8799-8a1decdb704c\") " pod="openstack/nova-cell0-5784-account-create-update-4gwt4" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.090270 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ngsz\" (UniqueName: \"kubernetes.io/projected/036fbace-3dcf-484d-958d-10970b9c7177-kube-api-access-7ngsz\") pod \"nova-cell1-db-create-jldcn\" (UID: \"036fbace-3dcf-484d-958d-10970b9c7177\") " pod="openstack/nova-cell1-db-create-jldcn" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.137614 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-9337-account-create-update-86n94"] Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.138864 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9337-account-create-update-86n94" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.140894 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.146315 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9337-account-create-update-86n94"] Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.170359 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7mk8\" (UniqueName: \"kubernetes.io/projected/797bc888-260e-418d-b8a7-e0ff2b48a9af-kube-api-access-w7mk8\") pod \"nova-cell1-9337-account-create-update-86n94\" (UID: \"797bc888-260e-418d-b8a7-e0ff2b48a9af\") " pod="openstack/nova-cell1-9337-account-create-update-86n94" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.170584 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/797bc888-260e-418d-b8a7-e0ff2b48a9af-operator-scripts\") pod \"nova-cell1-9337-account-create-update-86n94\" (UID: \"797bc888-260e-418d-b8a7-e0ff2b48a9af\") " pod="openstack/nova-cell1-9337-account-create-update-86n94" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.181914 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jldcn" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.252821 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5784-account-create-update-4gwt4" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.272804 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7mk8\" (UniqueName: \"kubernetes.io/projected/797bc888-260e-418d-b8a7-e0ff2b48a9af-kube-api-access-w7mk8\") pod \"nova-cell1-9337-account-create-update-86n94\" (UID: \"797bc888-260e-418d-b8a7-e0ff2b48a9af\") " pod="openstack/nova-cell1-9337-account-create-update-86n94" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.272868 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/797bc888-260e-418d-b8a7-e0ff2b48a9af-operator-scripts\") pod \"nova-cell1-9337-account-create-update-86n94\" (UID: \"797bc888-260e-418d-b8a7-e0ff2b48a9af\") " pod="openstack/nova-cell1-9337-account-create-update-86n94" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.273744 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/797bc888-260e-418d-b8a7-e0ff2b48a9af-operator-scripts\") pod \"nova-cell1-9337-account-create-update-86n94\" (UID: \"797bc888-260e-418d-b8a7-e0ff2b48a9af\") " pod="openstack/nova-cell1-9337-account-create-update-86n94" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.291251 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7mk8\" (UniqueName: \"kubernetes.io/projected/797bc888-260e-418d-b8a7-e0ff2b48a9af-kube-api-access-w7mk8\") pod \"nova-cell1-9337-account-create-update-86n94\" (UID: \"797bc888-260e-418d-b8a7-e0ff2b48a9af\") " pod="openstack/nova-cell1-9337-account-create-update-86n94" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.474788 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9337-account-create-update-86n94" Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.487774 4815 generic.go:334] "Generic (PLEG): container finished" podID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerID="0e3f2c6c27b9e6a889af97e0163bfa7460186ec10b5b75dc96cf6c63752c31b3" exitCode=0 Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.487819 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"178df77c-9b85-4f6e-8c47-44f47dc60383","Type":"ContainerDied","Data":"0e3f2c6c27b9e6a889af97e0163bfa7460186ec10b5b75dc96cf6c63752c31b3"} Feb 25 13:39:06 crc kubenswrapper[4815]: I0225 13:39:06.946803 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 25 13:39:07 crc kubenswrapper[4815]: I0225 13:39:07.215087 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.164:3000/\": dial tcp 10.217.0.164:3000: connect: connection refused" Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.155001 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.155250 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="96b0bb62-f8d6-4253-a70a-a95f071472df" containerName="glance-log" containerID="cri-o://bf90c6f53416fb5d0819b337fee413c8a1085ef1909bae7fb50aa793a4a122af" gracePeriod=30 Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.155385 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="96b0bb62-f8d6-4253-a70a-a95f071472df" containerName="glance-httpd" containerID="cri-o://9fbfb6af1d9b2620d8d826b615a233bc26850af324ad4bc7b503385bcffffea4" gracePeriod=30 Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.512977 4815 generic.go:334] "Generic (PLEG): container finished" podID="96b0bb62-f8d6-4253-a70a-a95f071472df" containerID="bf90c6f53416fb5d0819b337fee413c8a1085ef1909bae7fb50aa793a4a122af" exitCode=143 Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.513263 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"96b0bb62-f8d6-4253-a70a-a95f071472df","Type":"ContainerDied","Data":"bf90c6f53416fb5d0819b337fee413c8a1085ef1909bae7fb50aa793a4a122af"} Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.837188 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.925719 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/178df77c-9b85-4f6e-8c47-44f47dc60383-log-httpd\") pod \"178df77c-9b85-4f6e-8c47-44f47dc60383\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.925830 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-config-data\") pod \"178df77c-9b85-4f6e-8c47-44f47dc60383\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.925895 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-scripts\") pod \"178df77c-9b85-4f6e-8c47-44f47dc60383\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.925934 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-combined-ca-bundle\") pod \"178df77c-9b85-4f6e-8c47-44f47dc60383\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.926042 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-sg-core-conf-yaml\") pod \"178df77c-9b85-4f6e-8c47-44f47dc60383\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.926062 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhhwk\" (UniqueName: \"kubernetes.io/projected/178df77c-9b85-4f6e-8c47-44f47dc60383-kube-api-access-mhhwk\") pod \"178df77c-9b85-4f6e-8c47-44f47dc60383\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.926086 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/178df77c-9b85-4f6e-8c47-44f47dc60383-run-httpd\") pod \"178df77c-9b85-4f6e-8c47-44f47dc60383\" (UID: \"178df77c-9b85-4f6e-8c47-44f47dc60383\") " Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.930544 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/178df77c-9b85-4f6e-8c47-44f47dc60383-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "178df77c-9b85-4f6e-8c47-44f47dc60383" (UID: "178df77c-9b85-4f6e-8c47-44f47dc60383"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.930713 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/178df77c-9b85-4f6e-8c47-44f47dc60383-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "178df77c-9b85-4f6e-8c47-44f47dc60383" (UID: "178df77c-9b85-4f6e-8c47-44f47dc60383"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.931132 4815 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/178df77c-9b85-4f6e-8c47-44f47dc60383-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.931151 4815 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/178df77c-9b85-4f6e-8c47-44f47dc60383-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.939749 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/178df77c-9b85-4f6e-8c47-44f47dc60383-kube-api-access-mhhwk" (OuterVolumeSpecName: "kube-api-access-mhhwk") pod "178df77c-9b85-4f6e-8c47-44f47dc60383" (UID: "178df77c-9b85-4f6e-8c47-44f47dc60383"). InnerVolumeSpecName "kube-api-access-mhhwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.941697 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-scripts" (OuterVolumeSpecName: "scripts") pod "178df77c-9b85-4f6e-8c47-44f47dc60383" (UID: "178df77c-9b85-4f6e-8c47-44f47dc60383"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:08 crc kubenswrapper[4815]: I0225 13:39:08.961554 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "178df77c-9b85-4f6e-8c47-44f47dc60383" (UID: "178df77c-9b85-4f6e-8c47-44f47dc60383"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.008895 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "178df77c-9b85-4f6e-8c47-44f47dc60383" (UID: "178df77c-9b85-4f6e-8c47-44f47dc60383"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.026357 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-config-data" (OuterVolumeSpecName: "config-data") pod "178df77c-9b85-4f6e-8c47-44f47dc60383" (UID: "178df77c-9b85-4f6e-8c47-44f47dc60383"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.032431 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.032561 4815 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.032648 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhhwk\" (UniqueName: \"kubernetes.io/projected/178df77c-9b85-4f6e-8c47-44f47dc60383-kube-api-access-mhhwk\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.032747 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.032833 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/178df77c-9b85-4f6e-8c47-44f47dc60383-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.084429 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-gcl8c"] Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.182120 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-68d8cd5787-kmnfm"] Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.286659 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9337-account-create-update-86n94"] Feb 25 13:39:09 crc kubenswrapper[4815]: W0225 13:39:09.296071 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d139922_3032_4e41_978e_6bdd8ad8e0e1.slice/crio-8f857a850dad705548a169373eda10acbf020cccde9166dd54b0aee27a1f4bac WatchSource:0}: Error finding container 8f857a850dad705548a169373eda10acbf020cccde9166dd54b0aee27a1f4bac: Status 404 returned error can't find the container with id 8f857a850dad705548a169373eda10acbf020cccde9166dd54b0aee27a1f4bac Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.303950 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0d2e-account-create-update-c7zn7"] Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.474704 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-jldcn"] Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.487550 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-p95dh"] Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.538019 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.538025 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"178df77c-9b85-4f6e-8c47-44f47dc60383","Type":"ContainerDied","Data":"92e2f09c8e6644a7dc04b6f9a552463c154899a3cbf4b5e6a8c762cc72c72457"} Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.538205 4815 scope.go:117] "RemoveContainer" containerID="64fa5eafb58ae0796dbda8aa80d9b18239391d9bff07332f0ee6268798453139" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.543589 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0d2e-account-create-update-c7zn7" event={"ID":"7d139922-3032-4e41-978e-6bdd8ad8e0e1","Type":"ContainerStarted","Data":"8f857a850dad705548a169373eda10acbf020cccde9166dd54b0aee27a1f4bac"} Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.551317 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9337-account-create-update-86n94" event={"ID":"797bc888-260e-418d-b8a7-e0ff2b48a9af","Type":"ContainerStarted","Data":"3542a5ebd9f7a756f56492fdc9501e9752b8fb2aa089832e7171dab397695359"} Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.553212 4815 generic.go:334] "Generic (PLEG): container finished" podID="2da1710b-375f-4795-993d-730ccbd04ccc" containerID="87466353d055e76f7d2aac41f39bf94d60ddd3053fbcf2d908e16d25074f648c" exitCode=0 Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.553264 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-gcl8c" event={"ID":"2da1710b-375f-4795-993d-730ccbd04ccc","Type":"ContainerDied","Data":"87466353d055e76f7d2aac41f39bf94d60ddd3053fbcf2d908e16d25074f648c"} Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.553289 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-gcl8c" event={"ID":"2da1710b-375f-4795-993d-730ccbd04ccc","Type":"ContainerStarted","Data":"15333bbbb57b6948aa7d5aea336f3b75f4d7ef47f458f37363678d7bf052634b"} Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.554788 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-68d8cd5787-kmnfm" event={"ID":"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96","Type":"ContainerStarted","Data":"09fe15ba95b4111040856330080ebc5b4a826797aae177f90c03f1128253b32c"} Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.556074 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-p95dh" event={"ID":"486b281d-a6fc-4c69-860a-ad0d5480fc77","Type":"ContainerStarted","Data":"172408376f8c9024a2fdcf32c0e43e8dbb4127520e6d3dd561d6573f6c4d7aee"} Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.557754 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jldcn" event={"ID":"036fbace-3dcf-484d-958d-10970b9c7177","Type":"ContainerStarted","Data":"f6cc0804f03cfcb6f2364e9824aa24a627ec387e1c2935c672e2ea0e53cd6ec5"} Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.559321 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"3f935809-d8cf-4a5f-8d6b-05d7045b6bc8","Type":"ContainerStarted","Data":"0e05af2bc1d37e509141828182ef152bc52db95c159a464dc8f21377d5f42283"} Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.610425 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.3923797159999998 podStartE2EDuration="12.610408945s" podCreationTimestamp="2026-02-25 13:38:57 +0000 UTC" firstStartedPulling="2026-02-25 13:38:58.333781041 +0000 UTC m=+1096.134879095" lastFinishedPulling="2026-02-25 13:39:08.55181027 +0000 UTC m=+1106.352908324" observedRunningTime="2026-02-25 13:39:09.597344983 +0000 UTC m=+1107.398443047" watchObservedRunningTime="2026-02-25 13:39:09.610408945 +0000 UTC m=+1107.411506999" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.671675 4815 scope.go:117] "RemoveContainer" containerID="d7e37413d82a64dba8f1ff0c987f3cc5793ccc301f538991657226e2a5025542" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.696538 4815 scope.go:117] "RemoveContainer" containerID="0e3f2c6c27b9e6a889af97e0163bfa7460186ec10b5b75dc96cf6c63752c31b3" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.711615 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.735556 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-5784-account-create-update-4gwt4"] Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.741336 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.749568 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:09 crc kubenswrapper[4815]: E0225 13:39:09.749990 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerName="sg-core" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.750008 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerName="sg-core" Feb 25 13:39:09 crc kubenswrapper[4815]: E0225 13:39:09.750020 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerName="proxy-httpd" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.750027 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerName="proxy-httpd" Feb 25 13:39:09 crc kubenswrapper[4815]: E0225 13:39:09.750046 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerName="ceilometer-notification-agent" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.750052 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerName="ceilometer-notification-agent" Feb 25 13:39:09 crc kubenswrapper[4815]: E0225 13:39:09.750062 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerName="ceilometer-central-agent" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.750067 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerName="ceilometer-central-agent" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.750231 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerName="sg-core" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.750243 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerName="ceilometer-central-agent" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.750275 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerName="proxy-httpd" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.750285 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" containerName="ceilometer-notification-agent" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.751848 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.756714 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.758253 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.758639 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.763729 4815 scope.go:117] "RemoveContainer" containerID="dfe32d3f63f2e76c406dc47bcfa57eb2c27bc3a91d00b5753fa4909f18a475b3" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.854777 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.855496 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q27bb\" (UniqueName: \"kubernetes.io/projected/8bb4f144-223d-4fc5-992d-ac435ee0366b-kube-api-access-q27bb\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.855584 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.855781 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-scripts\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.855883 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-config-data\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.855971 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bb4f144-223d-4fc5-992d-ac435ee0366b-run-httpd\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.856021 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bb4f144-223d-4fc5-992d-ac435ee0366b-log-httpd\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.957557 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-scripts\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.957804 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-config-data\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.957930 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bb4f144-223d-4fc5-992d-ac435ee0366b-run-httpd\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.958037 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bb4f144-223d-4fc5-992d-ac435ee0366b-log-httpd\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.958254 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.958373 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q27bb\" (UniqueName: \"kubernetes.io/projected/8bb4f144-223d-4fc5-992d-ac435ee0366b-kube-api-access-q27bb\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.958569 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.960430 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bb4f144-223d-4fc5-992d-ac435ee0366b-run-httpd\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.966380 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.966536 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bb4f144-223d-4fc5-992d-ac435ee0366b-log-httpd\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.968254 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-scripts\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.968905 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-config-data\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.972351 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:09 crc kubenswrapper[4815]: I0225 13:39:09.988523 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q27bb\" (UniqueName: \"kubernetes.io/projected/8bb4f144-223d-4fc5-992d-ac435ee0366b-kube-api-access-q27bb\") pod \"ceilometer-0\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " pod="openstack/ceilometer-0" Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.025543 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.025748 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="532b0818-f59f-4b9d-a4fd-6a976bcdb886" containerName="glance-log" containerID="cri-o://7b7091745b0e6c271d6f260f9e7b2c007d9f3629133f78a3ad3bfd00245e119b" gracePeriod=30 Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.026019 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="532b0818-f59f-4b9d-a4fd-6a976bcdb886" containerName="glance-httpd" containerID="cri-o://a2b8f88ce679426f3e4de1e8142e92559d8f4ccd4926b3b118b36d8249c419a7" gracePeriod=30 Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.138990 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.496446 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.569839 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-68d8cd5787-kmnfm" event={"ID":"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96","Type":"ContainerStarted","Data":"beec124bd48f46c0f8c7ef1f3171cfd03cd7dbd635cb6cf4bc9665632c10b0b3"} Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.569893 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-68d8cd5787-kmnfm" event={"ID":"ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96","Type":"ContainerStarted","Data":"06fceec77ff9d359073dfe73078c2531b10bafc20b9087f47d52ff71e2ce9d7f"} Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.570111 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.570141 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.572520 4815 generic.go:334] "Generic (PLEG): container finished" podID="486b281d-a6fc-4c69-860a-ad0d5480fc77" containerID="1e214a4fc38a0f6ae8dfc700a53a44d035a833bae6de51f606dc41c4523c7914" exitCode=0 Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.572579 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-p95dh" event={"ID":"486b281d-a6fc-4c69-860a-ad0d5480fc77","Type":"ContainerDied","Data":"1e214a4fc38a0f6ae8dfc700a53a44d035a833bae6de51f606dc41c4523c7914"} Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.575010 4815 generic.go:334] "Generic (PLEG): container finished" podID="532b0818-f59f-4b9d-a4fd-6a976bcdb886" containerID="7b7091745b0e6c271d6f260f9e7b2c007d9f3629133f78a3ad3bfd00245e119b" exitCode=143 Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.575089 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"532b0818-f59f-4b9d-a4fd-6a976bcdb886","Type":"ContainerDied","Data":"7b7091745b0e6c271d6f260f9e7b2c007d9f3629133f78a3ad3bfd00245e119b"} Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.576831 4815 generic.go:334] "Generic (PLEG): container finished" podID="036fbace-3dcf-484d-958d-10970b9c7177" containerID="f48af9cccc860922855910f8e43bee585e03439414b4ad16b244780831db02b0" exitCode=0 Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.576997 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jldcn" event={"ID":"036fbace-3dcf-484d-958d-10970b9c7177","Type":"ContainerDied","Data":"f48af9cccc860922855910f8e43bee585e03439414b4ad16b244780831db02b0"} Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.579981 4815 generic.go:334] "Generic (PLEG): container finished" podID="7d139922-3032-4e41-978e-6bdd8ad8e0e1" containerID="584628a831021fe1c356deec621859ebe6086c452d288829b49333e39571b7b9" exitCode=0 Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.580085 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0d2e-account-create-update-c7zn7" event={"ID":"7d139922-3032-4e41-978e-6bdd8ad8e0e1","Type":"ContainerDied","Data":"584628a831021fe1c356deec621859ebe6086c452d288829b49333e39571b7b9"} Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.586080 4815 generic.go:334] "Generic (PLEG): container finished" podID="6648a53c-a316-42aa-8799-8a1decdb704c" containerID="0d7ba17420be972376accbbcf2ccefc5fffda46d83ab793c0d542847d6d77abf" exitCode=0 Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.586178 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5784-account-create-update-4gwt4" event={"ID":"6648a53c-a316-42aa-8799-8a1decdb704c","Type":"ContainerDied","Data":"0d7ba17420be972376accbbcf2ccefc5fffda46d83ab793c0d542847d6d77abf"} Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.586390 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5784-account-create-update-4gwt4" event={"ID":"6648a53c-a316-42aa-8799-8a1decdb704c","Type":"ContainerStarted","Data":"d5dc2cb28a4c3226fff55d129cf3c847c9c479b6cc00673fdc557de283bb87a2"} Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.588047 4815 generic.go:334] "Generic (PLEG): container finished" podID="797bc888-260e-418d-b8a7-e0ff2b48a9af" containerID="4ba8b1744c2f0247354c6cb552c320b03a8105617c24401eac941a74cdf5eb29" exitCode=0 Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.588074 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9337-account-create-update-86n94" event={"ID":"797bc888-260e-418d-b8a7-e0ff2b48a9af","Type":"ContainerDied","Data":"4ba8b1744c2f0247354c6cb552c320b03a8105617c24401eac941a74cdf5eb29"} Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.602718 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-68d8cd5787-kmnfm" podStartSLOduration=8.602696253 podStartE2EDuration="8.602696253s" podCreationTimestamp="2026-02-25 13:39:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:39:10.595365325 +0000 UTC m=+1108.396463379" watchObservedRunningTime="2026-02-25 13:39:10.602696253 +0000 UTC m=+1108.403794307" Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.662833 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.921167 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-gcl8c" Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.945973 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="178df77c-9b85-4f6e-8c47-44f47dc60383" path="/var/lib/kubelet/pods/178df77c-9b85-4f6e-8c47-44f47dc60383/volumes" Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.986298 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2da1710b-375f-4795-993d-730ccbd04ccc-operator-scripts\") pod \"2da1710b-375f-4795-993d-730ccbd04ccc\" (UID: \"2da1710b-375f-4795-993d-730ccbd04ccc\") " Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.986436 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjt46\" (UniqueName: \"kubernetes.io/projected/2da1710b-375f-4795-993d-730ccbd04ccc-kube-api-access-kjt46\") pod \"2da1710b-375f-4795-993d-730ccbd04ccc\" (UID: \"2da1710b-375f-4795-993d-730ccbd04ccc\") " Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.987053 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2da1710b-375f-4795-993d-730ccbd04ccc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2da1710b-375f-4795-993d-730ccbd04ccc" (UID: "2da1710b-375f-4795-993d-730ccbd04ccc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:39:10 crc kubenswrapper[4815]: I0225 13:39:10.994467 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2da1710b-375f-4795-993d-730ccbd04ccc-kube-api-access-kjt46" (OuterVolumeSpecName: "kube-api-access-kjt46") pod "2da1710b-375f-4795-993d-730ccbd04ccc" (UID: "2da1710b-375f-4795-993d-730ccbd04ccc"). InnerVolumeSpecName "kube-api-access-kjt46". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:39:11 crc kubenswrapper[4815]: I0225 13:39:11.088610 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjt46\" (UniqueName: \"kubernetes.io/projected/2da1710b-375f-4795-993d-730ccbd04ccc-kube-api-access-kjt46\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:11 crc kubenswrapper[4815]: I0225 13:39:11.088887 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2da1710b-375f-4795-993d-730ccbd04ccc-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:11 crc kubenswrapper[4815]: I0225 13:39:11.595964 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-gcl8c" event={"ID":"2da1710b-375f-4795-993d-730ccbd04ccc","Type":"ContainerDied","Data":"15333bbbb57b6948aa7d5aea336f3b75f4d7ef47f458f37363678d7bf052634b"} Feb 25 13:39:11 crc kubenswrapper[4815]: I0225 13:39:11.597284 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15333bbbb57b6948aa7d5aea336f3b75f4d7ef47f458f37363678d7bf052634b" Feb 25 13:39:11 crc kubenswrapper[4815]: I0225 13:39:11.596014 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-gcl8c" Feb 25 13:39:11 crc kubenswrapper[4815]: I0225 13:39:11.598493 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bb4f144-223d-4fc5-992d-ac435ee0366b","Type":"ContainerStarted","Data":"26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1"} Feb 25 13:39:11 crc kubenswrapper[4815]: I0225 13:39:11.598557 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bb4f144-223d-4fc5-992d-ac435ee0366b","Type":"ContainerStarted","Data":"f758e75e8f7c7620113ccdc0b1ea15c9b106df0ea6ab6d54aa71ebabcaabd4bd"} Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.034368 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-p95dh" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.105488 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjjz4\" (UniqueName: \"kubernetes.io/projected/486b281d-a6fc-4c69-860a-ad0d5480fc77-kube-api-access-vjjz4\") pod \"486b281d-a6fc-4c69-860a-ad0d5480fc77\" (UID: \"486b281d-a6fc-4c69-860a-ad0d5480fc77\") " Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.105727 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/486b281d-a6fc-4c69-860a-ad0d5480fc77-operator-scripts\") pod \"486b281d-a6fc-4c69-860a-ad0d5480fc77\" (UID: \"486b281d-a6fc-4c69-860a-ad0d5480fc77\") " Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.107478 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/486b281d-a6fc-4c69-860a-ad0d5480fc77-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "486b281d-a6fc-4c69-860a-ad0d5480fc77" (UID: "486b281d-a6fc-4c69-860a-ad0d5480fc77"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.129813 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/486b281d-a6fc-4c69-860a-ad0d5480fc77-kube-api-access-vjjz4" (OuterVolumeSpecName: "kube-api-access-vjjz4") pod "486b281d-a6fc-4c69-860a-ad0d5480fc77" (UID: "486b281d-a6fc-4c69-860a-ad0d5480fc77"). InnerVolumeSpecName "kube-api-access-vjjz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.167716 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9337-account-create-update-86n94" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.176713 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0d2e-account-create-update-c7zn7" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.200790 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5784-account-create-update-4gwt4" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.209849 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7mk8\" (UniqueName: \"kubernetes.io/projected/797bc888-260e-418d-b8a7-e0ff2b48a9af-kube-api-access-w7mk8\") pod \"797bc888-260e-418d-b8a7-e0ff2b48a9af\" (UID: \"797bc888-260e-418d-b8a7-e0ff2b48a9af\") " Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.210063 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/797bc888-260e-418d-b8a7-e0ff2b48a9af-operator-scripts\") pod \"797bc888-260e-418d-b8a7-e0ff2b48a9af\" (UID: \"797bc888-260e-418d-b8a7-e0ff2b48a9af\") " Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.210844 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/486b281d-a6fc-4c69-860a-ad0d5480fc77-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.210861 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjjz4\" (UniqueName: \"kubernetes.io/projected/486b281d-a6fc-4c69-860a-ad0d5480fc77-kube-api-access-vjjz4\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.220154 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/797bc888-260e-418d-b8a7-e0ff2b48a9af-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "797bc888-260e-418d-b8a7-e0ff2b48a9af" (UID: "797bc888-260e-418d-b8a7-e0ff2b48a9af"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.221023 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jldcn" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.224646 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/797bc888-260e-418d-b8a7-e0ff2b48a9af-kube-api-access-w7mk8" (OuterVolumeSpecName: "kube-api-access-w7mk8") pod "797bc888-260e-418d-b8a7-e0ff2b48a9af" (UID: "797bc888-260e-418d-b8a7-e0ff2b48a9af"). InnerVolumeSpecName "kube-api-access-w7mk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.312333 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d139922-3032-4e41-978e-6bdd8ad8e0e1-operator-scripts\") pod \"7d139922-3032-4e41-978e-6bdd8ad8e0e1\" (UID: \"7d139922-3032-4e41-978e-6bdd8ad8e0e1\") " Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.312380 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dkpp\" (UniqueName: \"kubernetes.io/projected/7d139922-3032-4e41-978e-6bdd8ad8e0e1-kube-api-access-9dkpp\") pod \"7d139922-3032-4e41-978e-6bdd8ad8e0e1\" (UID: \"7d139922-3032-4e41-978e-6bdd8ad8e0e1\") " Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.312406 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6648a53c-a316-42aa-8799-8a1decdb704c-operator-scripts\") pod \"6648a53c-a316-42aa-8799-8a1decdb704c\" (UID: \"6648a53c-a316-42aa-8799-8a1decdb704c\") " Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.312546 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/036fbace-3dcf-484d-958d-10970b9c7177-operator-scripts\") pod \"036fbace-3dcf-484d-958d-10970b9c7177\" (UID: \"036fbace-3dcf-484d-958d-10970b9c7177\") " Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.312619 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ngsz\" (UniqueName: \"kubernetes.io/projected/036fbace-3dcf-484d-958d-10970b9c7177-kube-api-access-7ngsz\") pod \"036fbace-3dcf-484d-958d-10970b9c7177\" (UID: \"036fbace-3dcf-484d-958d-10970b9c7177\") " Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.312697 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdnbv\" (UniqueName: \"kubernetes.io/projected/6648a53c-a316-42aa-8799-8a1decdb704c-kube-api-access-jdnbv\") pod \"6648a53c-a316-42aa-8799-8a1decdb704c\" (UID: \"6648a53c-a316-42aa-8799-8a1decdb704c\") " Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.312795 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d139922-3032-4e41-978e-6bdd8ad8e0e1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7d139922-3032-4e41-978e-6bdd8ad8e0e1" (UID: "7d139922-3032-4e41-978e-6bdd8ad8e0e1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.313035 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7mk8\" (UniqueName: \"kubernetes.io/projected/797bc888-260e-418d-b8a7-e0ff2b48a9af-kube-api-access-w7mk8\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.313053 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d139922-3032-4e41-978e-6bdd8ad8e0e1-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.313061 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/797bc888-260e-418d-b8a7-e0ff2b48a9af-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.313279 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/036fbace-3dcf-484d-958d-10970b9c7177-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "036fbace-3dcf-484d-958d-10970b9c7177" (UID: "036fbace-3dcf-484d-958d-10970b9c7177"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.313739 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6648a53c-a316-42aa-8799-8a1decdb704c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6648a53c-a316-42aa-8799-8a1decdb704c" (UID: "6648a53c-a316-42aa-8799-8a1decdb704c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.315575 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d139922-3032-4e41-978e-6bdd8ad8e0e1-kube-api-access-9dkpp" (OuterVolumeSpecName: "kube-api-access-9dkpp") pod "7d139922-3032-4e41-978e-6bdd8ad8e0e1" (UID: "7d139922-3032-4e41-978e-6bdd8ad8e0e1"). InnerVolumeSpecName "kube-api-access-9dkpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.316248 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6648a53c-a316-42aa-8799-8a1decdb704c-kube-api-access-jdnbv" (OuterVolumeSpecName: "kube-api-access-jdnbv") pod "6648a53c-a316-42aa-8799-8a1decdb704c" (UID: "6648a53c-a316-42aa-8799-8a1decdb704c"). InnerVolumeSpecName "kube-api-access-jdnbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.316901 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/036fbace-3dcf-484d-958d-10970b9c7177-kube-api-access-7ngsz" (OuterVolumeSpecName: "kube-api-access-7ngsz") pod "036fbace-3dcf-484d-958d-10970b9c7177" (UID: "036fbace-3dcf-484d-958d-10970b9c7177"). InnerVolumeSpecName "kube-api-access-7ngsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.414485 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/036fbace-3dcf-484d-958d-10970b9c7177-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.414532 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ngsz\" (UniqueName: \"kubernetes.io/projected/036fbace-3dcf-484d-958d-10970b9c7177-kube-api-access-7ngsz\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.414544 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdnbv\" (UniqueName: \"kubernetes.io/projected/6648a53c-a316-42aa-8799-8a1decdb704c-kube-api-access-jdnbv\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.414553 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dkpp\" (UniqueName: \"kubernetes.io/projected/7d139922-3032-4e41-978e-6bdd8ad8e0e1-kube-api-access-9dkpp\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.414562 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6648a53c-a316-42aa-8799-8a1decdb704c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.617054 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9337-account-create-update-86n94" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.618374 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9337-account-create-update-86n94" event={"ID":"797bc888-260e-418d-b8a7-e0ff2b48a9af","Type":"ContainerDied","Data":"3542a5ebd9f7a756f56492fdc9501e9752b8fb2aa089832e7171dab397695359"} Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.618421 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3542a5ebd9f7a756f56492fdc9501e9752b8fb2aa089832e7171dab397695359" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.626590 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bb4f144-223d-4fc5-992d-ac435ee0366b","Type":"ContainerStarted","Data":"be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59"} Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.637925 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-p95dh" event={"ID":"486b281d-a6fc-4c69-860a-ad0d5480fc77","Type":"ContainerDied","Data":"172408376f8c9024a2fdcf32c0e43e8dbb4127520e6d3dd561d6573f6c4d7aee"} Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.638252 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="172408376f8c9024a2fdcf32c0e43e8dbb4127520e6d3dd561d6573f6c4d7aee" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.638487 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-p95dh" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.642540 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jldcn" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.642494 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jldcn" event={"ID":"036fbace-3dcf-484d-958d-10970b9c7177","Type":"ContainerDied","Data":"f6cc0804f03cfcb6f2364e9824aa24a627ec387e1c2935c672e2ea0e53cd6ec5"} Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.642590 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6cc0804f03cfcb6f2364e9824aa24a627ec387e1c2935c672e2ea0e53cd6ec5" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.645949 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0d2e-account-create-update-c7zn7" event={"ID":"7d139922-3032-4e41-978e-6bdd8ad8e0e1","Type":"ContainerDied","Data":"8f857a850dad705548a169373eda10acbf020cccde9166dd54b0aee27a1f4bac"} Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.645986 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f857a850dad705548a169373eda10acbf020cccde9166dd54b0aee27a1f4bac" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.645963 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0d2e-account-create-update-c7zn7" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.647030 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5784-account-create-update-4gwt4" event={"ID":"6648a53c-a316-42aa-8799-8a1decdb704c","Type":"ContainerDied","Data":"d5dc2cb28a4c3226fff55d129cf3c847c9c479b6cc00673fdc557de283bb87a2"} Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.647054 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5dc2cb28a4c3226fff55d129cf3c847c9c479b6cc00673fdc557de283bb87a2" Feb 25 13:39:12 crc kubenswrapper[4815]: I0225 13:39:12.647087 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5784-account-create-update-4gwt4" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.074680 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.127382 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-config-data\") pod \"96b0bb62-f8d6-4253-a70a-a95f071472df\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.127434 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"96b0bb62-f8d6-4253-a70a-a95f071472df\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.127530 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/96b0bb62-f8d6-4253-a70a-a95f071472df-httpd-run\") pod \"96b0bb62-f8d6-4253-a70a-a95f071472df\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.127571 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-scripts\") pod \"96b0bb62-f8d6-4253-a70a-a95f071472df\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.127618 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpn4z\" (UniqueName: \"kubernetes.io/projected/96b0bb62-f8d6-4253-a70a-a95f071472df-kube-api-access-dpn4z\") pod \"96b0bb62-f8d6-4253-a70a-a95f071472df\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.127638 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-public-tls-certs\") pod \"96b0bb62-f8d6-4253-a70a-a95f071472df\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.127713 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-combined-ca-bundle\") pod \"96b0bb62-f8d6-4253-a70a-a95f071472df\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.127762 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96b0bb62-f8d6-4253-a70a-a95f071472df-logs\") pod \"96b0bb62-f8d6-4253-a70a-a95f071472df\" (UID: \"96b0bb62-f8d6-4253-a70a-a95f071472df\") " Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.129026 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96b0bb62-f8d6-4253-a70a-a95f071472df-logs" (OuterVolumeSpecName: "logs") pod "96b0bb62-f8d6-4253-a70a-a95f071472df" (UID: "96b0bb62-f8d6-4253-a70a-a95f071472df"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.134328 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96b0bb62-f8d6-4253-a70a-a95f071472df-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "96b0bb62-f8d6-4253-a70a-a95f071472df" (UID: "96b0bb62-f8d6-4253-a70a-a95f071472df"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.137442 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-scripts" (OuterVolumeSpecName: "scripts") pod "96b0bb62-f8d6-4253-a70a-a95f071472df" (UID: "96b0bb62-f8d6-4253-a70a-a95f071472df"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.137440 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "96b0bb62-f8d6-4253-a70a-a95f071472df" (UID: "96b0bb62-f8d6-4253-a70a-a95f071472df"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.140494 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b0bb62-f8d6-4253-a70a-a95f071472df-kube-api-access-dpn4z" (OuterVolumeSpecName: "kube-api-access-dpn4z") pod "96b0bb62-f8d6-4253-a70a-a95f071472df" (UID: "96b0bb62-f8d6-4253-a70a-a95f071472df"). InnerVolumeSpecName "kube-api-access-dpn4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.179127 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96b0bb62-f8d6-4253-a70a-a95f071472df" (UID: "96b0bb62-f8d6-4253-a70a-a95f071472df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.217657 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-config-data" (OuterVolumeSpecName: "config-data") pod "96b0bb62-f8d6-4253-a70a-a95f071472df" (UID: "96b0bb62-f8d6-4253-a70a-a95f071472df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.229794 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.229841 4815 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.229851 4815 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/96b0bb62-f8d6-4253-a70a-a95f071472df-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.229860 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.229869 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpn4z\" (UniqueName: \"kubernetes.io/projected/96b0bb62-f8d6-4253-a70a-a95f071472df-kube-api-access-dpn4z\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.229878 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.229887 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96b0bb62-f8d6-4253-a70a-a95f071472df-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.259922 4815 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.266621 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "96b0bb62-f8d6-4253-a70a-a95f071472df" (UID: "96b0bb62-f8d6-4253-a70a-a95f071472df"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.331750 4815 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/96b0bb62-f8d6-4253-a70a-a95f071472df-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.331775 4815 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.658214 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bb4f144-223d-4fc5-992d-ac435ee0366b","Type":"ContainerStarted","Data":"759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8"} Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.661442 4815 generic.go:334] "Generic (PLEG): container finished" podID="532b0818-f59f-4b9d-a4fd-6a976bcdb886" containerID="a2b8f88ce679426f3e4de1e8142e92559d8f4ccd4926b3b118b36d8249c419a7" exitCode=0 Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.661532 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"532b0818-f59f-4b9d-a4fd-6a976bcdb886","Type":"ContainerDied","Data":"a2b8f88ce679426f3e4de1e8142e92559d8f4ccd4926b3b118b36d8249c419a7"} Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.663480 4815 generic.go:334] "Generic (PLEG): container finished" podID="96b0bb62-f8d6-4253-a70a-a95f071472df" containerID="9fbfb6af1d9b2620d8d826b615a233bc26850af324ad4bc7b503385bcffffea4" exitCode=0 Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.663542 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"96b0bb62-f8d6-4253-a70a-a95f071472df","Type":"ContainerDied","Data":"9fbfb6af1d9b2620d8d826b615a233bc26850af324ad4bc7b503385bcffffea4"} Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.663568 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"96b0bb62-f8d6-4253-a70a-a95f071472df","Type":"ContainerDied","Data":"651b85a5d30918624c3825b1ff0fb632136bebcff4c904e1fdef8baed1491991"} Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.663584 4815 scope.go:117] "RemoveContainer" containerID="9fbfb6af1d9b2620d8d826b615a233bc26850af324ad4bc7b503385bcffffea4" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.663709 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.695476 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.708176 4815 scope.go:117] "RemoveContainer" containerID="bf90c6f53416fb5d0819b337fee413c8a1085ef1909bae7fb50aa793a4a122af" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.708800 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.725613 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:39:13 crc kubenswrapper[4815]: E0225 13:39:13.726886 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6648a53c-a316-42aa-8799-8a1decdb704c" containerName="mariadb-account-create-update" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.726916 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6648a53c-a316-42aa-8799-8a1decdb704c" containerName="mariadb-account-create-update" Feb 25 13:39:13 crc kubenswrapper[4815]: E0225 13:39:13.726932 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="486b281d-a6fc-4c69-860a-ad0d5480fc77" containerName="mariadb-database-create" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.726942 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="486b281d-a6fc-4c69-860a-ad0d5480fc77" containerName="mariadb-database-create" Feb 25 13:39:13 crc kubenswrapper[4815]: E0225 13:39:13.726958 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96b0bb62-f8d6-4253-a70a-a95f071472df" containerName="glance-log" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.726967 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="96b0bb62-f8d6-4253-a70a-a95f071472df" containerName="glance-log" Feb 25 13:39:13 crc kubenswrapper[4815]: E0225 13:39:13.726982 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="797bc888-260e-418d-b8a7-e0ff2b48a9af" containerName="mariadb-account-create-update" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.726990 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="797bc888-260e-418d-b8a7-e0ff2b48a9af" containerName="mariadb-account-create-update" Feb 25 13:39:13 crc kubenswrapper[4815]: E0225 13:39:13.727000 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96b0bb62-f8d6-4253-a70a-a95f071472df" containerName="glance-httpd" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.727010 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="96b0bb62-f8d6-4253-a70a-a95f071472df" containerName="glance-httpd" Feb 25 13:39:13 crc kubenswrapper[4815]: E0225 13:39:13.727030 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d139922-3032-4e41-978e-6bdd8ad8e0e1" containerName="mariadb-account-create-update" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.727038 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d139922-3032-4e41-978e-6bdd8ad8e0e1" containerName="mariadb-account-create-update" Feb 25 13:39:13 crc kubenswrapper[4815]: E0225 13:39:13.727089 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="036fbace-3dcf-484d-958d-10970b9c7177" containerName="mariadb-database-create" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.727102 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="036fbace-3dcf-484d-958d-10970b9c7177" containerName="mariadb-database-create" Feb 25 13:39:13 crc kubenswrapper[4815]: E0225 13:39:13.727114 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2da1710b-375f-4795-993d-730ccbd04ccc" containerName="mariadb-database-create" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.727123 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2da1710b-375f-4795-993d-730ccbd04ccc" containerName="mariadb-database-create" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.727378 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="6648a53c-a316-42aa-8799-8a1decdb704c" containerName="mariadb-account-create-update" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.727407 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="96b0bb62-f8d6-4253-a70a-a95f071472df" containerName="glance-log" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.727436 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="486b281d-a6fc-4c69-860a-ad0d5480fc77" containerName="mariadb-database-create" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.727447 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="2da1710b-375f-4795-993d-730ccbd04ccc" containerName="mariadb-database-create" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.727463 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="036fbace-3dcf-484d-958d-10970b9c7177" containerName="mariadb-database-create" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.727481 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="797bc888-260e-418d-b8a7-e0ff2b48a9af" containerName="mariadb-account-create-update" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.727500 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d139922-3032-4e41-978e-6bdd8ad8e0e1" containerName="mariadb-account-create-update" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.727529 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="96b0bb62-f8d6-4253-a70a-a95f071472df" containerName="glance-httpd" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.730076 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.732113 4815 scope.go:117] "RemoveContainer" containerID="9fbfb6af1d9b2620d8d826b615a233bc26850af324ad4bc7b503385bcffffea4" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.732389 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.732531 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 25 13:39:13 crc kubenswrapper[4815]: E0225 13:39:13.733093 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fbfb6af1d9b2620d8d826b615a233bc26850af324ad4bc7b503385bcffffea4\": container with ID starting with 9fbfb6af1d9b2620d8d826b615a233bc26850af324ad4bc7b503385bcffffea4 not found: ID does not exist" containerID="9fbfb6af1d9b2620d8d826b615a233bc26850af324ad4bc7b503385bcffffea4" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.733119 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fbfb6af1d9b2620d8d826b615a233bc26850af324ad4bc7b503385bcffffea4"} err="failed to get container status \"9fbfb6af1d9b2620d8d826b615a233bc26850af324ad4bc7b503385bcffffea4\": rpc error: code = NotFound desc = could not find container \"9fbfb6af1d9b2620d8d826b615a233bc26850af324ad4bc7b503385bcffffea4\": container with ID starting with 9fbfb6af1d9b2620d8d826b615a233bc26850af324ad4bc7b503385bcffffea4 not found: ID does not exist" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.733168 4815 scope.go:117] "RemoveContainer" containerID="bf90c6f53416fb5d0819b337fee413c8a1085ef1909bae7fb50aa793a4a122af" Feb 25 13:39:13 crc kubenswrapper[4815]: E0225 13:39:13.734001 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf90c6f53416fb5d0819b337fee413c8a1085ef1909bae7fb50aa793a4a122af\": container with ID starting with bf90c6f53416fb5d0819b337fee413c8a1085ef1909bae7fb50aa793a4a122af not found: ID does not exist" containerID="bf90c6f53416fb5d0819b337fee413c8a1085ef1909bae7fb50aa793a4a122af" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.734091 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf90c6f53416fb5d0819b337fee413c8a1085ef1909bae7fb50aa793a4a122af"} err="failed to get container status \"bf90c6f53416fb5d0819b337fee413c8a1085ef1909bae7fb50aa793a4a122af\": rpc error: code = NotFound desc = could not find container \"bf90c6f53416fb5d0819b337fee413c8a1085ef1909bae7fb50aa793a4a122af\": container with ID starting with bf90c6f53416fb5d0819b337fee413c8a1085ef1909bae7fb50aa793a4a122af not found: ID does not exist" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.748580 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.837344 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6497d456d-s5g59" podUID="5d076fbb-4499-456f-82ee-211a8909fc99" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.837653 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.838870 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-logs\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.838904 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.838938 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-scripts\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.838962 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-config-data\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.838993 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.839014 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf2pz\" (UniqueName: \"kubernetes.io/projected/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-kube-api-access-tf2pz\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.839048 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.839077 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.941003 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-logs\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.941063 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.941117 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-scripts\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.941161 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-config-data\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.941217 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.941250 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf2pz\" (UniqueName: \"kubernetes.io/projected/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-kube-api-access-tf2pz\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.941311 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.941367 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.941398 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.941547 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-logs\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.941769 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.953160 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.953530 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.955914 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-config-data\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.957654 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-scripts\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.959722 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf2pz\" (UniqueName: \"kubernetes.io/projected/f30dde44-ecd6-4908-b18e-1f9ba7f43fd7-kube-api-access-tf2pz\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:13 crc kubenswrapper[4815]: I0225 13:39:13.991451 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7\") " pod="openstack/glance-default-external-api-0" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.021190 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-68dffb47c7-mfx8z" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.081190 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6fdb68948-zzdwv"] Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.081406 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6fdb68948-zzdwv" podUID="b02ef5af-64a9-4fbc-b312-30ac0ffab7ee" containerName="neutron-api" containerID="cri-o://ef9ce2c46a0c5e9b4cda1d459ecff2c694b65b93036e13dbb2d1c30c9858cde7" gracePeriod=30 Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.081471 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6fdb68948-zzdwv" podUID="b02ef5af-64a9-4fbc-b312-30ac0ffab7ee" containerName="neutron-httpd" containerID="cri-o://e8b05302d38f26d33ea76e9b4625bfd62a5ec72ad4bcb7929790aa9bc15eab5d" gracePeriod=30 Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.111031 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.242912 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.346926 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-internal-tls-certs\") pod \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.347218 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-config-data\") pod \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.347237 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/532b0818-f59f-4b9d-a4fd-6a976bcdb886-logs\") pod \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.347268 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-scripts\") pod \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.347289 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.347305 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/532b0818-f59f-4b9d-a4fd-6a976bcdb886-httpd-run\") pod \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.347376 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gshpz\" (UniqueName: \"kubernetes.io/projected/532b0818-f59f-4b9d-a4fd-6a976bcdb886-kube-api-access-gshpz\") pod \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.347411 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-combined-ca-bundle\") pod \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\" (UID: \"532b0818-f59f-4b9d-a4fd-6a976bcdb886\") " Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.347990 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/532b0818-f59f-4b9d-a4fd-6a976bcdb886-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "532b0818-f59f-4b9d-a4fd-6a976bcdb886" (UID: "532b0818-f59f-4b9d-a4fd-6a976bcdb886"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.348532 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/532b0818-f59f-4b9d-a4fd-6a976bcdb886-logs" (OuterVolumeSpecName: "logs") pod "532b0818-f59f-4b9d-a4fd-6a976bcdb886" (UID: "532b0818-f59f-4b9d-a4fd-6a976bcdb886"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.357673 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "532b0818-f59f-4b9d-a4fd-6a976bcdb886" (UID: "532b0818-f59f-4b9d-a4fd-6a976bcdb886"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.357718 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-scripts" (OuterVolumeSpecName: "scripts") pod "532b0818-f59f-4b9d-a4fd-6a976bcdb886" (UID: "532b0818-f59f-4b9d-a4fd-6a976bcdb886"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.358852 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/532b0818-f59f-4b9d-a4fd-6a976bcdb886-kube-api-access-gshpz" (OuterVolumeSpecName: "kube-api-access-gshpz") pod "532b0818-f59f-4b9d-a4fd-6a976bcdb886" (UID: "532b0818-f59f-4b9d-a4fd-6a976bcdb886"). InnerVolumeSpecName "kube-api-access-gshpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.423328 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "532b0818-f59f-4b9d-a4fd-6a976bcdb886" (UID: "532b0818-f59f-4b9d-a4fd-6a976bcdb886"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.428039 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "532b0818-f59f-4b9d-a4fd-6a976bcdb886" (UID: "532b0818-f59f-4b9d-a4fd-6a976bcdb886"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.450797 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gshpz\" (UniqueName: \"kubernetes.io/projected/532b0818-f59f-4b9d-a4fd-6a976bcdb886-kube-api-access-gshpz\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.450821 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.450831 4815 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.450839 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/532b0818-f59f-4b9d-a4fd-6a976bcdb886-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.450848 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.450868 4815 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.450877 4815 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/532b0818-f59f-4b9d-a4fd-6a976bcdb886-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.460149 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-config-data" (OuterVolumeSpecName: "config-data") pod "532b0818-f59f-4b9d-a4fd-6a976bcdb886" (UID: "532b0818-f59f-4b9d-a4fd-6a976bcdb886"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.491789 4815 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.552881 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/532b0818-f59f-4b9d-a4fd-6a976bcdb886-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.552920 4815 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.678364 4815 generic.go:334] "Generic (PLEG): container finished" podID="b02ef5af-64a9-4fbc-b312-30ac0ffab7ee" containerID="e8b05302d38f26d33ea76e9b4625bfd62a5ec72ad4bcb7929790aa9bc15eab5d" exitCode=0 Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.678407 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fdb68948-zzdwv" event={"ID":"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee","Type":"ContainerDied","Data":"e8b05302d38f26d33ea76e9b4625bfd62a5ec72ad4bcb7929790aa9bc15eab5d"} Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.682480 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"532b0818-f59f-4b9d-a4fd-6a976bcdb886","Type":"ContainerDied","Data":"17b60340b74377f53425b3acfc7f303c43fcbbba844febba81bcb8cf2bc6485e"} Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.682529 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.682556 4815 scope.go:117] "RemoveContainer" containerID="a2b8f88ce679426f3e4de1e8142e92559d8f4ccd4926b3b118b36d8249c419a7" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.744780 4815 scope.go:117] "RemoveContainer" containerID="7b7091745b0e6c271d6f260f9e7b2c007d9f3629133f78a3ad3bfd00245e119b" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.757557 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.772613 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.798759 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 13:39:14 crc kubenswrapper[4815]: E0225 13:39:14.799216 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="532b0818-f59f-4b9d-a4fd-6a976bcdb886" containerName="glance-log" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.799238 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="532b0818-f59f-4b9d-a4fd-6a976bcdb886" containerName="glance-log" Feb 25 13:39:14 crc kubenswrapper[4815]: E0225 13:39:14.799254 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="532b0818-f59f-4b9d-a4fd-6a976bcdb886" containerName="glance-httpd" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.799261 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="532b0818-f59f-4b9d-a4fd-6a976bcdb886" containerName="glance-httpd" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.799445 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="532b0818-f59f-4b9d-a4fd-6a976bcdb886" containerName="glance-log" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.799464 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="532b0818-f59f-4b9d-a4fd-6a976bcdb886" containerName="glance-httpd" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.800414 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.805529 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.805719 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.834682 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.872979 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.945079 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="532b0818-f59f-4b9d-a4fd-6a976bcdb886" path="/var/lib/kubelet/pods/532b0818-f59f-4b9d-a4fd-6a976bcdb886/volumes" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.945762 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b0bb62-f8d6-4253-a70a-a95f071472df" path="/var/lib/kubelet/pods/96b0bb62-f8d6-4253-a70a-a95f071472df/volumes" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.968230 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsljd\" (UniqueName: \"kubernetes.io/projected/17cf15ab-9607-484e-b043-fffcd641dc99-kube-api-access-lsljd\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.968288 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.968346 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17cf15ab-9607-484e-b043-fffcd641dc99-scripts\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.968639 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17cf15ab-9607-484e-b043-fffcd641dc99-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.968682 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/17cf15ab-9607-484e-b043-fffcd641dc99-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.968768 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17cf15ab-9607-484e-b043-fffcd641dc99-config-data\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.968823 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17cf15ab-9607-484e-b043-fffcd641dc99-logs\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:14 crc kubenswrapper[4815]: I0225 13:39:14.968860 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17cf15ab-9607-484e-b043-fffcd641dc99-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.070749 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsljd\" (UniqueName: \"kubernetes.io/projected/17cf15ab-9607-484e-b043-fffcd641dc99-kube-api-access-lsljd\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.070835 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.070925 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17cf15ab-9607-484e-b043-fffcd641dc99-scripts\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.071009 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17cf15ab-9607-484e-b043-fffcd641dc99-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.071029 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/17cf15ab-9607-484e-b043-fffcd641dc99-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.071086 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17cf15ab-9607-484e-b043-fffcd641dc99-config-data\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.071114 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17cf15ab-9607-484e-b043-fffcd641dc99-logs\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.071153 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17cf15ab-9607-484e-b043-fffcd641dc99-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.071223 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.071800 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17cf15ab-9607-484e-b043-fffcd641dc99-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.071915 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17cf15ab-9607-484e-b043-fffcd641dc99-logs\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.075685 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/17cf15ab-9607-484e-b043-fffcd641dc99-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.076281 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17cf15ab-9607-484e-b043-fffcd641dc99-config-data\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.078212 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17cf15ab-9607-484e-b043-fffcd641dc99-scripts\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.090047 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17cf15ab-9607-484e-b043-fffcd641dc99-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.093308 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsljd\" (UniqueName: \"kubernetes.io/projected/17cf15ab-9607-484e-b043-fffcd641dc99-kube-api-access-lsljd\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.107754 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"17cf15ab-9607-484e-b043-fffcd641dc99\") " pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.127745 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.699926 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.724481 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7","Type":"ContainerStarted","Data":"043e09f6b1ce1217b207010030e80602e38ddd3226ce1ad5f87665b9ca97bde4"} Feb 25 13:39:15 crc kubenswrapper[4815]: I0225 13:39:15.724610 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7","Type":"ContainerStarted","Data":"2b281b125f67f63a2f6e4fe7aa6bcc8e5dfed0bdac61085f089b766082e5fa81"} Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.208013 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6t9rm"] Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.210699 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6t9rm" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.213226 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.213484 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-lstls" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.216007 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.261517 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6t9rm"] Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.304272 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-scripts\") pod \"nova-cell0-conductor-db-sync-6t9rm\" (UID: \"e2580f26-4938-40dc-8a0f-044232a70469\") " pod="openstack/nova-cell0-conductor-db-sync-6t9rm" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.304316 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6t9rm\" (UID: \"e2580f26-4938-40dc-8a0f-044232a70469\") " pod="openstack/nova-cell0-conductor-db-sync-6t9rm" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.304377 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vrmx\" (UniqueName: \"kubernetes.io/projected/e2580f26-4938-40dc-8a0f-044232a70469-kube-api-access-7vrmx\") pod \"nova-cell0-conductor-db-sync-6t9rm\" (UID: \"e2580f26-4938-40dc-8a0f-044232a70469\") " pod="openstack/nova-cell0-conductor-db-sync-6t9rm" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.304416 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-config-data\") pod \"nova-cell0-conductor-db-sync-6t9rm\" (UID: \"e2580f26-4938-40dc-8a0f-044232a70469\") " pod="openstack/nova-cell0-conductor-db-sync-6t9rm" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.406427 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-scripts\") pod \"nova-cell0-conductor-db-sync-6t9rm\" (UID: \"e2580f26-4938-40dc-8a0f-044232a70469\") " pod="openstack/nova-cell0-conductor-db-sync-6t9rm" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.406477 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6t9rm\" (UID: \"e2580f26-4938-40dc-8a0f-044232a70469\") " pod="openstack/nova-cell0-conductor-db-sync-6t9rm" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.406575 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vrmx\" (UniqueName: \"kubernetes.io/projected/e2580f26-4938-40dc-8a0f-044232a70469-kube-api-access-7vrmx\") pod \"nova-cell0-conductor-db-sync-6t9rm\" (UID: \"e2580f26-4938-40dc-8a0f-044232a70469\") " pod="openstack/nova-cell0-conductor-db-sync-6t9rm" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.406891 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-config-data\") pod \"nova-cell0-conductor-db-sync-6t9rm\" (UID: \"e2580f26-4938-40dc-8a0f-044232a70469\") " pod="openstack/nova-cell0-conductor-db-sync-6t9rm" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.419247 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6t9rm\" (UID: \"e2580f26-4938-40dc-8a0f-044232a70469\") " pod="openstack/nova-cell0-conductor-db-sync-6t9rm" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.420824 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-config-data\") pod \"nova-cell0-conductor-db-sync-6t9rm\" (UID: \"e2580f26-4938-40dc-8a0f-044232a70469\") " pod="openstack/nova-cell0-conductor-db-sync-6t9rm" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.421156 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-scripts\") pod \"nova-cell0-conductor-db-sync-6t9rm\" (UID: \"e2580f26-4938-40dc-8a0f-044232a70469\") " pod="openstack/nova-cell0-conductor-db-sync-6t9rm" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.435467 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vrmx\" (UniqueName: \"kubernetes.io/projected/e2580f26-4938-40dc-8a0f-044232a70469-kube-api-access-7vrmx\") pod \"nova-cell0-conductor-db-sync-6t9rm\" (UID: \"e2580f26-4938-40dc-8a0f-044232a70469\") " pod="openstack/nova-cell0-conductor-db-sync-6t9rm" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.529052 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6t9rm" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.737652 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bb4f144-223d-4fc5-992d-ac435ee0366b","Type":"ContainerStarted","Data":"91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0"} Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.737797 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerName="ceilometer-central-agent" containerID="cri-o://26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1" gracePeriod=30 Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.737872 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.738177 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerName="proxy-httpd" containerID="cri-o://91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0" gracePeriod=30 Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.738236 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerName="sg-core" containerID="cri-o://759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8" gracePeriod=30 Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.738267 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerName="ceilometer-notification-agent" containerID="cri-o://be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59" gracePeriod=30 Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.744607 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f30dde44-ecd6-4908-b18e-1f9ba7f43fd7","Type":"ContainerStarted","Data":"46f7268b0eba02194029126f26d2d73ae669ddc424c0a6530a404f5acc0a0665"} Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.748676 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"17cf15ab-9607-484e-b043-fffcd641dc99","Type":"ContainerStarted","Data":"6ae834d0c9bab411a39fca20e8651c66438c79882b90ccc0e59fb2cd669cca90"} Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.748992 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"17cf15ab-9607-484e-b043-fffcd641dc99","Type":"ContainerStarted","Data":"b2ddcf3386ae026a756097bbc2b77ef9a0a74351155dfffa60b20ffa8a02b769"} Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.769766 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.812491671 podStartE2EDuration="7.769744512s" podCreationTimestamp="2026-02-25 13:39:09 +0000 UTC" firstStartedPulling="2026-02-25 13:39:10.658675836 +0000 UTC m=+1108.459773890" lastFinishedPulling="2026-02-25 13:39:15.615928687 +0000 UTC m=+1113.417026731" observedRunningTime="2026-02-25 13:39:16.762374993 +0000 UTC m=+1114.563473047" watchObservedRunningTime="2026-02-25 13:39:16.769744512 +0000 UTC m=+1114.570842556" Feb 25 13:39:16 crc kubenswrapper[4815]: I0225 13:39:16.798789 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.798770899 podStartE2EDuration="3.798770899s" podCreationTimestamp="2026-02-25 13:39:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:39:16.78687653 +0000 UTC m=+1114.587974584" watchObservedRunningTime="2026-02-25 13:39:16.798770899 +0000 UTC m=+1114.599868953" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.040866 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6t9rm"] Feb 25 13:39:17 crc kubenswrapper[4815]: W0225 13:39:17.054589 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2580f26_4938_40dc_8a0f_044232a70469.slice/crio-c0ce3ac48de315f34c630b709d75944a991e1a5ccde0433004e6b765023ba1d5 WatchSource:0}: Error finding container c0ce3ac48de315f34c630b709d75944a991e1a5ccde0433004e6b765023ba1d5: Status 404 returned error can't find the container with id c0ce3ac48de315f34c630b709d75944a991e1a5ccde0433004e6b765023ba1d5 Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.701379 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.710656 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.773982 4815 generic.go:334] "Generic (PLEG): container finished" podID="b02ef5af-64a9-4fbc-b312-30ac0ffab7ee" containerID="ef9ce2c46a0c5e9b4cda1d459ecff2c694b65b93036e13dbb2d1c30c9858cde7" exitCode=0 Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.774049 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6fdb68948-zzdwv" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.774776 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fdb68948-zzdwv" event={"ID":"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee","Type":"ContainerDied","Data":"ef9ce2c46a0c5e9b4cda1d459ecff2c694b65b93036e13dbb2d1c30c9858cde7"} Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.774894 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fdb68948-zzdwv" event={"ID":"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee","Type":"ContainerDied","Data":"cff5a29db301d89f37170debc4530bf24f0e2220e4646766d268e8fca776c2ed"} Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.774922 4815 scope.go:117] "RemoveContainer" containerID="e8b05302d38f26d33ea76e9b4625bfd62a5ec72ad4bcb7929790aa9bc15eab5d" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.781431 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"17cf15ab-9607-484e-b043-fffcd641dc99","Type":"ContainerStarted","Data":"17b8ca7052b4193781a7d134153668233807950449b7f6cb9c10e7ac46076435"} Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.788899 4815 generic.go:334] "Generic (PLEG): container finished" podID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerID="91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0" exitCode=0 Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.788930 4815 generic.go:334] "Generic (PLEG): container finished" podID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerID="759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8" exitCode=2 Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.788938 4815 generic.go:334] "Generic (PLEG): container finished" podID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerID="be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59" exitCode=0 Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.788946 4815 generic.go:334] "Generic (PLEG): container finished" podID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerID="26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1" exitCode=0 Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.788978 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bb4f144-223d-4fc5-992d-ac435ee0366b","Type":"ContainerDied","Data":"91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0"} Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.789001 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bb4f144-223d-4fc5-992d-ac435ee0366b","Type":"ContainerDied","Data":"759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8"} Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.789011 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bb4f144-223d-4fc5-992d-ac435ee0366b","Type":"ContainerDied","Data":"be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59"} Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.789021 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bb4f144-223d-4fc5-992d-ac435ee0366b","Type":"ContainerDied","Data":"26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1"} Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.789031 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8bb4f144-223d-4fc5-992d-ac435ee0366b","Type":"ContainerDied","Data":"f758e75e8f7c7620113ccdc0b1ea15c9b106df0ea6ab6d54aa71ebabcaabd4bd"} Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.789088 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.791174 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6t9rm" event={"ID":"e2580f26-4938-40dc-8a0f-044232a70469","Type":"ContainerStarted","Data":"c0ce3ac48de315f34c630b709d75944a991e1a5ccde0433004e6b765023ba1d5"} Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.804911 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.8048948 podStartE2EDuration="3.8048948s" podCreationTimestamp="2026-02-25 13:39:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:39:17.80242053 +0000 UTC m=+1115.603518594" watchObservedRunningTime="2026-02-25 13:39:17.8048948 +0000 UTC m=+1115.605992854" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.819096 4815 scope.go:117] "RemoveContainer" containerID="ef9ce2c46a0c5e9b4cda1d459ecff2c694b65b93036e13dbb2d1c30c9858cde7" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.829782 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q27bb\" (UniqueName: \"kubernetes.io/projected/8bb4f144-223d-4fc5-992d-ac435ee0366b-kube-api-access-q27bb\") pod \"8bb4f144-223d-4fc5-992d-ac435ee0366b\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.829887 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bb4f144-223d-4fc5-992d-ac435ee0366b-log-httpd\") pod \"8bb4f144-223d-4fc5-992d-ac435ee0366b\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.829921 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxsqh\" (UniqueName: \"kubernetes.io/projected/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-kube-api-access-zxsqh\") pod \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.829937 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-config-data\") pod \"8bb4f144-223d-4fc5-992d-ac435ee0366b\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.829984 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-httpd-config\") pod \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.830021 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bb4f144-223d-4fc5-992d-ac435ee0366b-run-httpd\") pod \"8bb4f144-223d-4fc5-992d-ac435ee0366b\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.830071 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-ovndb-tls-certs\") pod \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.830086 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-config\") pod \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.830109 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-scripts\") pod \"8bb4f144-223d-4fc5-992d-ac435ee0366b\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.830134 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-combined-ca-bundle\") pod \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\" (UID: \"b02ef5af-64a9-4fbc-b312-30ac0ffab7ee\") " Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.830195 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-combined-ca-bundle\") pod \"8bb4f144-223d-4fc5-992d-ac435ee0366b\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.830221 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-sg-core-conf-yaml\") pod \"8bb4f144-223d-4fc5-992d-ac435ee0366b\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.833855 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bb4f144-223d-4fc5-992d-ac435ee0366b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8bb4f144-223d-4fc5-992d-ac435ee0366b" (UID: "8bb4f144-223d-4fc5-992d-ac435ee0366b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.837097 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bb4f144-223d-4fc5-992d-ac435ee0366b-kube-api-access-q27bb" (OuterVolumeSpecName: "kube-api-access-q27bb") pod "8bb4f144-223d-4fc5-992d-ac435ee0366b" (UID: "8bb4f144-223d-4fc5-992d-ac435ee0366b"). InnerVolumeSpecName "kube-api-access-q27bb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.839870 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bb4f144-223d-4fc5-992d-ac435ee0366b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8bb4f144-223d-4fc5-992d-ac435ee0366b" (UID: "8bb4f144-223d-4fc5-992d-ac435ee0366b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.842623 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-kube-api-access-zxsqh" (OuterVolumeSpecName: "kube-api-access-zxsqh") pod "b02ef5af-64a9-4fbc-b312-30ac0ffab7ee" (UID: "b02ef5af-64a9-4fbc-b312-30ac0ffab7ee"). InnerVolumeSpecName "kube-api-access-zxsqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.843020 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-scripts" (OuterVolumeSpecName: "scripts") pod "8bb4f144-223d-4fc5-992d-ac435ee0366b" (UID: "8bb4f144-223d-4fc5-992d-ac435ee0366b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.843680 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "b02ef5af-64a9-4fbc-b312-30ac0ffab7ee" (UID: "b02ef5af-64a9-4fbc-b312-30ac0ffab7ee"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.852487 4815 scope.go:117] "RemoveContainer" containerID="e8b05302d38f26d33ea76e9b4625bfd62a5ec72ad4bcb7929790aa9bc15eab5d" Feb 25 13:39:17 crc kubenswrapper[4815]: E0225 13:39:17.853761 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8b05302d38f26d33ea76e9b4625bfd62a5ec72ad4bcb7929790aa9bc15eab5d\": container with ID starting with e8b05302d38f26d33ea76e9b4625bfd62a5ec72ad4bcb7929790aa9bc15eab5d not found: ID does not exist" containerID="e8b05302d38f26d33ea76e9b4625bfd62a5ec72ad4bcb7929790aa9bc15eab5d" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.853804 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8b05302d38f26d33ea76e9b4625bfd62a5ec72ad4bcb7929790aa9bc15eab5d"} err="failed to get container status \"e8b05302d38f26d33ea76e9b4625bfd62a5ec72ad4bcb7929790aa9bc15eab5d\": rpc error: code = NotFound desc = could not find container \"e8b05302d38f26d33ea76e9b4625bfd62a5ec72ad4bcb7929790aa9bc15eab5d\": container with ID starting with e8b05302d38f26d33ea76e9b4625bfd62a5ec72ad4bcb7929790aa9bc15eab5d not found: ID does not exist" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.853831 4815 scope.go:117] "RemoveContainer" containerID="ef9ce2c46a0c5e9b4cda1d459ecff2c694b65b93036e13dbb2d1c30c9858cde7" Feb 25 13:39:17 crc kubenswrapper[4815]: E0225 13:39:17.855017 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef9ce2c46a0c5e9b4cda1d459ecff2c694b65b93036e13dbb2d1c30c9858cde7\": container with ID starting with ef9ce2c46a0c5e9b4cda1d459ecff2c694b65b93036e13dbb2d1c30c9858cde7 not found: ID does not exist" containerID="ef9ce2c46a0c5e9b4cda1d459ecff2c694b65b93036e13dbb2d1c30c9858cde7" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.855341 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef9ce2c46a0c5e9b4cda1d459ecff2c694b65b93036e13dbb2d1c30c9858cde7"} err="failed to get container status \"ef9ce2c46a0c5e9b4cda1d459ecff2c694b65b93036e13dbb2d1c30c9858cde7\": rpc error: code = NotFound desc = could not find container \"ef9ce2c46a0c5e9b4cda1d459ecff2c694b65b93036e13dbb2d1c30c9858cde7\": container with ID starting with ef9ce2c46a0c5e9b4cda1d459ecff2c694b65b93036e13dbb2d1c30c9858cde7 not found: ID does not exist" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.855972 4815 scope.go:117] "RemoveContainer" containerID="91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.877025 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.889969 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-68d8cd5787-kmnfm" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.890417 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8bb4f144-223d-4fc5-992d-ac435ee0366b" (UID: "8bb4f144-223d-4fc5-992d-ac435ee0366b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.932679 4815 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bb4f144-223d-4fc5-992d-ac435ee0366b-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.932895 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxsqh\" (UniqueName: \"kubernetes.io/projected/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-kube-api-access-zxsqh\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.932952 4815 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.933012 4815 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8bb4f144-223d-4fc5-992d-ac435ee0366b-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.933064 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.933121 4815 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.933179 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q27bb\" (UniqueName: \"kubernetes.io/projected/8bb4f144-223d-4fc5-992d-ac435ee0366b-kube-api-access-q27bb\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.949679 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-config" (OuterVolumeSpecName: "config") pod "b02ef5af-64a9-4fbc-b312-30ac0ffab7ee" (UID: "b02ef5af-64a9-4fbc-b312-30ac0ffab7ee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.957668 4815 scope.go:117] "RemoveContainer" containerID="759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.963976 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "b02ef5af-64a9-4fbc-b312-30ac0ffab7ee" (UID: "b02ef5af-64a9-4fbc-b312-30ac0ffab7ee"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.968638 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b02ef5af-64a9-4fbc-b312-30ac0ffab7ee" (UID: "b02ef5af-64a9-4fbc-b312-30ac0ffab7ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:17 crc kubenswrapper[4815]: I0225 13:39:17.987024 4815 scope.go:117] "RemoveContainer" containerID="be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59" Feb 25 13:39:17 crc kubenswrapper[4815]: E0225 13:39:17.994488 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-config-data podName:8bb4f144-223d-4fc5-992d-ac435ee0366b nodeName:}" failed. No retries permitted until 2026-02-25 13:39:18.494451545 +0000 UTC m=+1116.295549599 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-config-data") pod "8bb4f144-223d-4fc5-992d-ac435ee0366b" (UID: "8bb4f144-223d-4fc5-992d-ac435ee0366b") : error deleting /var/lib/kubelet/pods/8bb4f144-223d-4fc5-992d-ac435ee0366b/volume-subpaths: remove /var/lib/kubelet/pods/8bb4f144-223d-4fc5-992d-ac435ee0366b/volume-subpaths: no such file or directory Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.000751 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8bb4f144-223d-4fc5-992d-ac435ee0366b" (UID: "8bb4f144-223d-4fc5-992d-ac435ee0366b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.035274 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.035316 4815 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.035328 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.035342 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.043026 4815 scope.go:117] "RemoveContainer" containerID="26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.077592 4815 scope.go:117] "RemoveContainer" containerID="91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0" Feb 25 13:39:18 crc kubenswrapper[4815]: E0225 13:39:18.078001 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0\": container with ID starting with 91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0 not found: ID does not exist" containerID="91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.078058 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0"} err="failed to get container status \"91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0\": rpc error: code = NotFound desc = could not find container \"91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0\": container with ID starting with 91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0 not found: ID does not exist" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.078089 4815 scope.go:117] "RemoveContainer" containerID="759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8" Feb 25 13:39:18 crc kubenswrapper[4815]: E0225 13:39:18.078350 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8\": container with ID starting with 759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8 not found: ID does not exist" containerID="759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.078397 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8"} err="failed to get container status \"759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8\": rpc error: code = NotFound desc = could not find container \"759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8\": container with ID starting with 759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8 not found: ID does not exist" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.078419 4815 scope.go:117] "RemoveContainer" containerID="be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59" Feb 25 13:39:18 crc kubenswrapper[4815]: E0225 13:39:18.078658 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59\": container with ID starting with be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59 not found: ID does not exist" containerID="be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.078707 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59"} err="failed to get container status \"be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59\": rpc error: code = NotFound desc = could not find container \"be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59\": container with ID starting with be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59 not found: ID does not exist" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.078726 4815 scope.go:117] "RemoveContainer" containerID="26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1" Feb 25 13:39:18 crc kubenswrapper[4815]: E0225 13:39:18.078913 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1\": container with ID starting with 26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1 not found: ID does not exist" containerID="26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.078938 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1"} err="failed to get container status \"26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1\": rpc error: code = NotFound desc = could not find container \"26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1\": container with ID starting with 26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1 not found: ID does not exist" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.078953 4815 scope.go:117] "RemoveContainer" containerID="91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.079133 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0"} err="failed to get container status \"91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0\": rpc error: code = NotFound desc = could not find container \"91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0\": container with ID starting with 91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0 not found: ID does not exist" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.079161 4815 scope.go:117] "RemoveContainer" containerID="759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.079417 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8"} err="failed to get container status \"759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8\": rpc error: code = NotFound desc = could not find container \"759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8\": container with ID starting with 759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8 not found: ID does not exist" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.079460 4815 scope.go:117] "RemoveContainer" containerID="be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.080247 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59"} err="failed to get container status \"be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59\": rpc error: code = NotFound desc = could not find container \"be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59\": container with ID starting with be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59 not found: ID does not exist" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.080283 4815 scope.go:117] "RemoveContainer" containerID="26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.080612 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1"} err="failed to get container status \"26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1\": rpc error: code = NotFound desc = could not find container \"26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1\": container with ID starting with 26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1 not found: ID does not exist" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.080700 4815 scope.go:117] "RemoveContainer" containerID="91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.081088 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0"} err="failed to get container status \"91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0\": rpc error: code = NotFound desc = could not find container \"91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0\": container with ID starting with 91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0 not found: ID does not exist" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.081129 4815 scope.go:117] "RemoveContainer" containerID="759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.081449 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8"} err="failed to get container status \"759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8\": rpc error: code = NotFound desc = could not find container \"759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8\": container with ID starting with 759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8 not found: ID does not exist" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.081468 4815 scope.go:117] "RemoveContainer" containerID="be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.081708 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59"} err="failed to get container status \"be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59\": rpc error: code = NotFound desc = could not find container \"be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59\": container with ID starting with be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59 not found: ID does not exist" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.081729 4815 scope.go:117] "RemoveContainer" containerID="26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.083260 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1"} err="failed to get container status \"26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1\": rpc error: code = NotFound desc = could not find container \"26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1\": container with ID starting with 26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1 not found: ID does not exist" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.083282 4815 scope.go:117] "RemoveContainer" containerID="91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.083532 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0"} err="failed to get container status \"91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0\": rpc error: code = NotFound desc = could not find container \"91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0\": container with ID starting with 91ff6e4cbaa24fd4b4e7d61afe068a9ba401e6a380c54b79a205f5b6977014e0 not found: ID does not exist" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.083553 4815 scope.go:117] "RemoveContainer" containerID="759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.083834 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8"} err="failed to get container status \"759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8\": rpc error: code = NotFound desc = could not find container \"759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8\": container with ID starting with 759ce5cc6477b2143f413d8a276b0f774b711be707c6195caed0fb237ed057a8 not found: ID does not exist" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.083874 4815 scope.go:117] "RemoveContainer" containerID="be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.084090 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59"} err="failed to get container status \"be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59\": rpc error: code = NotFound desc = could not find container \"be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59\": container with ID starting with be587c4bdb44323b94d4e3b6b6df5530bcc0c46f846e118e8d1e8d15fce89f59 not found: ID does not exist" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.084109 4815 scope.go:117] "RemoveContainer" containerID="26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.084270 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1"} err="failed to get container status \"26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1\": rpc error: code = NotFound desc = could not find container \"26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1\": container with ID starting with 26edc47f557c4a7a6027892f5e127bc3de6d39f94ba7c04838948bfff44bc7d1 not found: ID does not exist" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.106657 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6fdb68948-zzdwv"] Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.114732 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6fdb68948-zzdwv"] Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.548605 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-config-data\") pod \"8bb4f144-223d-4fc5-992d-ac435ee0366b\" (UID: \"8bb4f144-223d-4fc5-992d-ac435ee0366b\") " Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.553123 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-config-data" (OuterVolumeSpecName: "config-data") pod "8bb4f144-223d-4fc5-992d-ac435ee0366b" (UID: "8bb4f144-223d-4fc5-992d-ac435ee0366b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.652825 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb4f144-223d-4fc5-992d-ac435ee0366b-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.732579 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.746387 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.755396 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:18 crc kubenswrapper[4815]: E0225 13:39:18.755775 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b02ef5af-64a9-4fbc-b312-30ac0ffab7ee" containerName="neutron-httpd" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.755791 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b02ef5af-64a9-4fbc-b312-30ac0ffab7ee" containerName="neutron-httpd" Feb 25 13:39:18 crc kubenswrapper[4815]: E0225 13:39:18.755806 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerName="ceilometer-notification-agent" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.755813 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerName="ceilometer-notification-agent" Feb 25 13:39:18 crc kubenswrapper[4815]: E0225 13:39:18.755836 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerName="proxy-httpd" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.755842 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerName="proxy-httpd" Feb 25 13:39:18 crc kubenswrapper[4815]: E0225 13:39:18.755852 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b02ef5af-64a9-4fbc-b312-30ac0ffab7ee" containerName="neutron-api" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.755858 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b02ef5af-64a9-4fbc-b312-30ac0ffab7ee" containerName="neutron-api" Feb 25 13:39:18 crc kubenswrapper[4815]: E0225 13:39:18.755875 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerName="ceilometer-central-agent" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.755881 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerName="ceilometer-central-agent" Feb 25 13:39:18 crc kubenswrapper[4815]: E0225 13:39:18.755893 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerName="sg-core" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.755899 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerName="sg-core" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.756049 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerName="sg-core" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.756067 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b02ef5af-64a9-4fbc-b312-30ac0ffab7ee" containerName="neutron-api" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.756083 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerName="proxy-httpd" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.756094 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b02ef5af-64a9-4fbc-b312-30ac0ffab7ee" containerName="neutron-httpd" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.756102 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerName="ceilometer-central-agent" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.756120 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bb4f144-223d-4fc5-992d-ac435ee0366b" containerName="ceilometer-notification-agent" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.758766 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.764039 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.764192 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.772040 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.857272 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.857401 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/647b5718-0f7b-42e3-aadc-209a60dfc054-log-httpd\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.857487 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-config-data\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.857525 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfk5t\" (UniqueName: \"kubernetes.io/projected/647b5718-0f7b-42e3-aadc-209a60dfc054-kube-api-access-hfk5t\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.857558 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-scripts\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.857590 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.857621 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/647b5718-0f7b-42e3-aadc-209a60dfc054-run-httpd\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.950727 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bb4f144-223d-4fc5-992d-ac435ee0366b" path="/var/lib/kubelet/pods/8bb4f144-223d-4fc5-992d-ac435ee0366b/volumes" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.952278 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b02ef5af-64a9-4fbc-b312-30ac0ffab7ee" path="/var/lib/kubelet/pods/b02ef5af-64a9-4fbc-b312-30ac0ffab7ee/volumes" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.960261 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.960822 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/647b5718-0f7b-42e3-aadc-209a60dfc054-run-httpd\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.961100 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.961150 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/647b5718-0f7b-42e3-aadc-209a60dfc054-run-httpd\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.962533 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/647b5718-0f7b-42e3-aadc-209a60dfc054-log-httpd\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.963245 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/647b5718-0f7b-42e3-aadc-209a60dfc054-log-httpd\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.964059 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-config-data\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.964094 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfk5t\" (UniqueName: \"kubernetes.io/projected/647b5718-0f7b-42e3-aadc-209a60dfc054-kube-api-access-hfk5t\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.964181 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-scripts\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.965195 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.969788 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-scripts\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.969797 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.972163 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-config-data\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:18 crc kubenswrapper[4815]: I0225 13:39:18.979179 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfk5t\" (UniqueName: \"kubernetes.io/projected/647b5718-0f7b-42e3-aadc-209a60dfc054-kube-api-access-hfk5t\") pod \"ceilometer-0\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " pod="openstack/ceilometer-0" Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.086217 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:39:19 crc kubenswrapper[4815]: W0225 13:39:19.261921 4815 container.go:586] Failed to update stats for container "/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d076fbb_4499_456f_82ee_211a8909fc99.slice/crio-79c17dbf7bf2ac9651809fd2e4d5f718ceebe6086b34ddd6ad5c389628898018": error while statting cgroup v2: [unable to parse /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d076fbb_4499_456f_82ee_211a8909fc99.slice/crio-79c17dbf7bf2ac9651809fd2e4d5f718ceebe6086b34ddd6ad5c389628898018/memory.stat: read /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d076fbb_4499_456f_82ee_211a8909fc99.slice/crio-79c17dbf7bf2ac9651809fd2e4d5f718ceebe6086b34ddd6ad5c389628898018/memory.stat: no such device], continuing to push stats Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.532318 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.588242 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:19 crc kubenswrapper[4815]: W0225 13:39:19.590627 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod647b5718_0f7b_42e3_aadc_209a60dfc054.slice/crio-dc7c167b6baa30bda9c2aaf44b0223a266efcab5740f2701b132450ecfface9c WatchSource:0}: Error finding container dc7c167b6baa30bda9c2aaf44b0223a266efcab5740f2701b132450ecfface9c: Status 404 returned error can't find the container with id dc7c167b6baa30bda9c2aaf44b0223a266efcab5740f2701b132450ecfface9c Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.685807 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.782952 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d076fbb-4499-456f-82ee-211a8909fc99-config-data\") pod \"5d076fbb-4499-456f-82ee-211a8909fc99\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.783267 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d076fbb-4499-456f-82ee-211a8909fc99-logs\") pod \"5d076fbb-4499-456f-82ee-211a8909fc99\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.783293 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-horizon-tls-certs\") pod \"5d076fbb-4499-456f-82ee-211a8909fc99\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.783343 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d076fbb-4499-456f-82ee-211a8909fc99-scripts\") pod \"5d076fbb-4499-456f-82ee-211a8909fc99\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.783385 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4jm8\" (UniqueName: \"kubernetes.io/projected/5d076fbb-4499-456f-82ee-211a8909fc99-kube-api-access-r4jm8\") pod \"5d076fbb-4499-456f-82ee-211a8909fc99\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.783581 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-combined-ca-bundle\") pod \"5d076fbb-4499-456f-82ee-211a8909fc99\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.783630 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-horizon-secret-key\") pod \"5d076fbb-4499-456f-82ee-211a8909fc99\" (UID: \"5d076fbb-4499-456f-82ee-211a8909fc99\") " Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.784179 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d076fbb-4499-456f-82ee-211a8909fc99-logs" (OuterVolumeSpecName: "logs") pod "5d076fbb-4499-456f-82ee-211a8909fc99" (UID: "5d076fbb-4499-456f-82ee-211a8909fc99"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.792179 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "5d076fbb-4499-456f-82ee-211a8909fc99" (UID: "5d076fbb-4499-456f-82ee-211a8909fc99"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.815885 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d076fbb-4499-456f-82ee-211a8909fc99-kube-api-access-r4jm8" (OuterVolumeSpecName: "kube-api-access-r4jm8") pod "5d076fbb-4499-456f-82ee-211a8909fc99" (UID: "5d076fbb-4499-456f-82ee-211a8909fc99"). InnerVolumeSpecName "kube-api-access-r4jm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.816550 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d076fbb-4499-456f-82ee-211a8909fc99-config-data" (OuterVolumeSpecName: "config-data") pod "5d076fbb-4499-456f-82ee-211a8909fc99" (UID: "5d076fbb-4499-456f-82ee-211a8909fc99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.830203 4815 generic.go:334] "Generic (PLEG): container finished" podID="5d076fbb-4499-456f-82ee-211a8909fc99" containerID="37b004a1e59b2810a4a4e2a9e741191118c1925114ad66b49a3a6e92a2d9e11e" exitCode=137 Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.830279 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6497d456d-s5g59" Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.830284 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6497d456d-s5g59" event={"ID":"5d076fbb-4499-456f-82ee-211a8909fc99","Type":"ContainerDied","Data":"37b004a1e59b2810a4a4e2a9e741191118c1925114ad66b49a3a6e92a2d9e11e"} Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.830634 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6497d456d-s5g59" event={"ID":"5d076fbb-4499-456f-82ee-211a8909fc99","Type":"ContainerDied","Data":"79c17dbf7bf2ac9651809fd2e4d5f718ceebe6086b34ddd6ad5c389628898018"} Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.830680 4815 scope.go:117] "RemoveContainer" containerID="9fba6bb1e21ffc87dd876586c19b77148678fd408e930bd6088e1ed1c188d47d" Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.834538 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"647b5718-0f7b-42e3-aadc-209a60dfc054","Type":"ContainerStarted","Data":"dc7c167b6baa30bda9c2aaf44b0223a266efcab5740f2701b132450ecfface9c"} Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.841730 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d076fbb-4499-456f-82ee-211a8909fc99" (UID: "5d076fbb-4499-456f-82ee-211a8909fc99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.858486 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d076fbb-4499-456f-82ee-211a8909fc99-scripts" (OuterVolumeSpecName: "scripts") pod "5d076fbb-4499-456f-82ee-211a8909fc99" (UID: "5d076fbb-4499-456f-82ee-211a8909fc99"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.865586 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "5d076fbb-4499-456f-82ee-211a8909fc99" (UID: "5d076fbb-4499-456f-82ee-211a8909fc99"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.889603 4815 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.889644 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d076fbb-4499-456f-82ee-211a8909fc99-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.889656 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d076fbb-4499-456f-82ee-211a8909fc99-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.889701 4815 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.889715 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d076fbb-4499-456f-82ee-211a8909fc99-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.889727 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4jm8\" (UniqueName: \"kubernetes.io/projected/5d076fbb-4499-456f-82ee-211a8909fc99-kube-api-access-r4jm8\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:19 crc kubenswrapper[4815]: I0225 13:39:19.889740 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d076fbb-4499-456f-82ee-211a8909fc99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:20 crc kubenswrapper[4815]: I0225 13:39:20.084057 4815 scope.go:117] "RemoveContainer" containerID="37b004a1e59b2810a4a4e2a9e741191118c1925114ad66b49a3a6e92a2d9e11e" Feb 25 13:39:20 crc kubenswrapper[4815]: I0225 13:39:20.116829 4815 scope.go:117] "RemoveContainer" containerID="9fba6bb1e21ffc87dd876586c19b77148678fd408e930bd6088e1ed1c188d47d" Feb 25 13:39:20 crc kubenswrapper[4815]: E0225 13:39:20.117351 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fba6bb1e21ffc87dd876586c19b77148678fd408e930bd6088e1ed1c188d47d\": container with ID starting with 9fba6bb1e21ffc87dd876586c19b77148678fd408e930bd6088e1ed1c188d47d not found: ID does not exist" containerID="9fba6bb1e21ffc87dd876586c19b77148678fd408e930bd6088e1ed1c188d47d" Feb 25 13:39:20 crc kubenswrapper[4815]: I0225 13:39:20.117388 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fba6bb1e21ffc87dd876586c19b77148678fd408e930bd6088e1ed1c188d47d"} err="failed to get container status \"9fba6bb1e21ffc87dd876586c19b77148678fd408e930bd6088e1ed1c188d47d\": rpc error: code = NotFound desc = could not find container \"9fba6bb1e21ffc87dd876586c19b77148678fd408e930bd6088e1ed1c188d47d\": container with ID starting with 9fba6bb1e21ffc87dd876586c19b77148678fd408e930bd6088e1ed1c188d47d not found: ID does not exist" Feb 25 13:39:20 crc kubenswrapper[4815]: I0225 13:39:20.117413 4815 scope.go:117] "RemoveContainer" containerID="37b004a1e59b2810a4a4e2a9e741191118c1925114ad66b49a3a6e92a2d9e11e" Feb 25 13:39:20 crc kubenswrapper[4815]: E0225 13:39:20.117766 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37b004a1e59b2810a4a4e2a9e741191118c1925114ad66b49a3a6e92a2d9e11e\": container with ID starting with 37b004a1e59b2810a4a4e2a9e741191118c1925114ad66b49a3a6e92a2d9e11e not found: ID does not exist" containerID="37b004a1e59b2810a4a4e2a9e741191118c1925114ad66b49a3a6e92a2d9e11e" Feb 25 13:39:20 crc kubenswrapper[4815]: I0225 13:39:20.117816 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37b004a1e59b2810a4a4e2a9e741191118c1925114ad66b49a3a6e92a2d9e11e"} err="failed to get container status \"37b004a1e59b2810a4a4e2a9e741191118c1925114ad66b49a3a6e92a2d9e11e\": rpc error: code = NotFound desc = could not find container \"37b004a1e59b2810a4a4e2a9e741191118c1925114ad66b49a3a6e92a2d9e11e\": container with ID starting with 37b004a1e59b2810a4a4e2a9e741191118c1925114ad66b49a3a6e92a2d9e11e not found: ID does not exist" Feb 25 13:39:20 crc kubenswrapper[4815]: I0225 13:39:20.175842 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6497d456d-s5g59"] Feb 25 13:39:20 crc kubenswrapper[4815]: I0225 13:39:20.184685 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6497d456d-s5g59"] Feb 25 13:39:20 crc kubenswrapper[4815]: I0225 13:39:20.850466 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"647b5718-0f7b-42e3-aadc-209a60dfc054","Type":"ContainerStarted","Data":"bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19"} Feb 25 13:39:20 crc kubenswrapper[4815]: I0225 13:39:20.947962 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d076fbb-4499-456f-82ee-211a8909fc99" path="/var/lib/kubelet/pods/5d076fbb-4499-456f-82ee-211a8909fc99/volumes" Feb 25 13:39:21 crc kubenswrapper[4815]: I0225 13:39:21.342933 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:39:21 crc kubenswrapper[4815]: I0225 13:39:21.343214 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:39:24 crc kubenswrapper[4815]: I0225 13:39:24.111980 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 25 13:39:24 crc kubenswrapper[4815]: I0225 13:39:24.112345 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 25 13:39:24 crc kubenswrapper[4815]: I0225 13:39:24.147482 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 25 13:39:24 crc kubenswrapper[4815]: I0225 13:39:24.158774 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 25 13:39:24 crc kubenswrapper[4815]: I0225 13:39:24.891975 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 25 13:39:24 crc kubenswrapper[4815]: I0225 13:39:24.892211 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 25 13:39:25 crc kubenswrapper[4815]: I0225 13:39:25.127944 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 25 13:39:25 crc kubenswrapper[4815]: I0225 13:39:25.128241 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 25 13:39:25 crc kubenswrapper[4815]: I0225 13:39:25.167606 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 25 13:39:25 crc kubenswrapper[4815]: I0225 13:39:25.185346 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 25 13:39:25 crc kubenswrapper[4815]: I0225 13:39:25.903797 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"647b5718-0f7b-42e3-aadc-209a60dfc054","Type":"ContainerStarted","Data":"b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242"} Feb 25 13:39:25 crc kubenswrapper[4815]: I0225 13:39:25.904393 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"647b5718-0f7b-42e3-aadc-209a60dfc054","Type":"ContainerStarted","Data":"19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887"} Feb 25 13:39:25 crc kubenswrapper[4815]: I0225 13:39:25.906305 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6t9rm" event={"ID":"e2580f26-4938-40dc-8a0f-044232a70469","Type":"ContainerStarted","Data":"ce07eb1dae824de5b0f51fe31b9a37f69fba77d6ee95300dc5ae30b254ebf417"} Feb 25 13:39:25 crc kubenswrapper[4815]: I0225 13:39:25.907113 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 25 13:39:25 crc kubenswrapper[4815]: I0225 13:39:25.907212 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 25 13:39:25 crc kubenswrapper[4815]: I0225 13:39:25.922843 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-6t9rm" podStartSLOduration=1.81398677 podStartE2EDuration="9.922801567s" podCreationTimestamp="2026-02-25 13:39:16 +0000 UTC" firstStartedPulling="2026-02-25 13:39:17.059235381 +0000 UTC m=+1114.860333435" lastFinishedPulling="2026-02-25 13:39:25.168050178 +0000 UTC m=+1122.969148232" observedRunningTime="2026-02-25 13:39:25.920524763 +0000 UTC m=+1123.721622817" watchObservedRunningTime="2026-02-25 13:39:25.922801567 +0000 UTC m=+1123.723899631" Feb 25 13:39:26 crc kubenswrapper[4815]: I0225 13:39:26.643256 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 25 13:39:26 crc kubenswrapper[4815]: I0225 13:39:26.834145 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 25 13:39:27 crc kubenswrapper[4815]: I0225 13:39:27.826487 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 25 13:39:27 crc kubenswrapper[4815]: I0225 13:39:27.826922 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 25 13:39:28 crc kubenswrapper[4815]: I0225 13:39:28.934432 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"647b5718-0f7b-42e3-aadc-209a60dfc054","Type":"ContainerStarted","Data":"7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28"} Feb 25 13:39:28 crc kubenswrapper[4815]: I0225 13:39:28.934983 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerName="ceilometer-central-agent" containerID="cri-o://bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19" gracePeriod=30 Feb 25 13:39:28 crc kubenswrapper[4815]: I0225 13:39:28.935355 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 13:39:28 crc kubenswrapper[4815]: I0225 13:39:28.935723 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerName="proxy-httpd" containerID="cri-o://7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28" gracePeriod=30 Feb 25 13:39:28 crc kubenswrapper[4815]: I0225 13:39:28.935781 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerName="sg-core" containerID="cri-o://b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242" gracePeriod=30 Feb 25 13:39:28 crc kubenswrapper[4815]: I0225 13:39:28.935834 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerName="ceilometer-notification-agent" containerID="cri-o://19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887" gracePeriod=30 Feb 25 13:39:28 crc kubenswrapper[4815]: I0225 13:39:28.978556 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.1566049720000002 podStartE2EDuration="10.978535176s" podCreationTimestamp="2026-02-25 13:39:18 +0000 UTC" firstStartedPulling="2026-02-25 13:39:19.594177389 +0000 UTC m=+1117.395275443" lastFinishedPulling="2026-02-25 13:39:27.416107593 +0000 UTC m=+1125.217205647" observedRunningTime="2026-02-25 13:39:28.974758129 +0000 UTC m=+1126.775856183" watchObservedRunningTime="2026-02-25 13:39:28.978535176 +0000 UTC m=+1126.779633230" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.727428 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.796266 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-sg-core-conf-yaml\") pod \"647b5718-0f7b-42e3-aadc-209a60dfc054\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.796394 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/647b5718-0f7b-42e3-aadc-209a60dfc054-run-httpd\") pod \"647b5718-0f7b-42e3-aadc-209a60dfc054\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.796417 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-combined-ca-bundle\") pod \"647b5718-0f7b-42e3-aadc-209a60dfc054\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.796441 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfk5t\" (UniqueName: \"kubernetes.io/projected/647b5718-0f7b-42e3-aadc-209a60dfc054-kube-api-access-hfk5t\") pod \"647b5718-0f7b-42e3-aadc-209a60dfc054\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.796467 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-config-data\") pod \"647b5718-0f7b-42e3-aadc-209a60dfc054\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.796522 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/647b5718-0f7b-42e3-aadc-209a60dfc054-log-httpd\") pod \"647b5718-0f7b-42e3-aadc-209a60dfc054\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.796564 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-scripts\") pod \"647b5718-0f7b-42e3-aadc-209a60dfc054\" (UID: \"647b5718-0f7b-42e3-aadc-209a60dfc054\") " Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.796975 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/647b5718-0f7b-42e3-aadc-209a60dfc054-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "647b5718-0f7b-42e3-aadc-209a60dfc054" (UID: "647b5718-0f7b-42e3-aadc-209a60dfc054"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.797785 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/647b5718-0f7b-42e3-aadc-209a60dfc054-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "647b5718-0f7b-42e3-aadc-209a60dfc054" (UID: "647b5718-0f7b-42e3-aadc-209a60dfc054"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.804608 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-scripts" (OuterVolumeSpecName: "scripts") pod "647b5718-0f7b-42e3-aadc-209a60dfc054" (UID: "647b5718-0f7b-42e3-aadc-209a60dfc054"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.804692 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/647b5718-0f7b-42e3-aadc-209a60dfc054-kube-api-access-hfk5t" (OuterVolumeSpecName: "kube-api-access-hfk5t") pod "647b5718-0f7b-42e3-aadc-209a60dfc054" (UID: "647b5718-0f7b-42e3-aadc-209a60dfc054"). InnerVolumeSpecName "kube-api-access-hfk5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.827142 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "647b5718-0f7b-42e3-aadc-209a60dfc054" (UID: "647b5718-0f7b-42e3-aadc-209a60dfc054"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.868609 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "647b5718-0f7b-42e3-aadc-209a60dfc054" (UID: "647b5718-0f7b-42e3-aadc-209a60dfc054"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.898066 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-config-data" (OuterVolumeSpecName: "config-data") pod "647b5718-0f7b-42e3-aadc-209a60dfc054" (UID: "647b5718-0f7b-42e3-aadc-209a60dfc054"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.898911 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.898944 4815 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.898959 4815 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/647b5718-0f7b-42e3-aadc-209a60dfc054-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.898970 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.898982 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfk5t\" (UniqueName: \"kubernetes.io/projected/647b5718-0f7b-42e3-aadc-209a60dfc054-kube-api-access-hfk5t\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.898994 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/647b5718-0f7b-42e3-aadc-209a60dfc054-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.899004 4815 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/647b5718-0f7b-42e3-aadc-209a60dfc054-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.943138 4815 generic.go:334] "Generic (PLEG): container finished" podID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerID="7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28" exitCode=0 Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.943168 4815 generic.go:334] "Generic (PLEG): container finished" podID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerID="b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242" exitCode=2 Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.943177 4815 generic.go:334] "Generic (PLEG): container finished" podID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerID="19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887" exitCode=0 Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.943185 4815 generic.go:334] "Generic (PLEG): container finished" podID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerID="bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19" exitCode=0 Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.943203 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"647b5718-0f7b-42e3-aadc-209a60dfc054","Type":"ContainerDied","Data":"7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28"} Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.943226 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"647b5718-0f7b-42e3-aadc-209a60dfc054","Type":"ContainerDied","Data":"b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242"} Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.943237 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"647b5718-0f7b-42e3-aadc-209a60dfc054","Type":"ContainerDied","Data":"19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887"} Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.943245 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"647b5718-0f7b-42e3-aadc-209a60dfc054","Type":"ContainerDied","Data":"bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19"} Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.943253 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"647b5718-0f7b-42e3-aadc-209a60dfc054","Type":"ContainerDied","Data":"dc7c167b6baa30bda9c2aaf44b0223a266efcab5740f2701b132450ecfface9c"} Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.943267 4815 scope.go:117] "RemoveContainer" containerID="7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.943378 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.981578 4815 scope.go:117] "RemoveContainer" containerID="b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.982469 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.991432 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.999181 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:29 crc kubenswrapper[4815]: E0225 13:39:29.999684 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerName="sg-core" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.999705 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerName="sg-core" Feb 25 13:39:29 crc kubenswrapper[4815]: E0225 13:39:29.999727 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerName="proxy-httpd" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.999735 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerName="proxy-httpd" Feb 25 13:39:29 crc kubenswrapper[4815]: E0225 13:39:29.999761 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d076fbb-4499-456f-82ee-211a8909fc99" containerName="horizon" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.999769 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d076fbb-4499-456f-82ee-211a8909fc99" containerName="horizon" Feb 25 13:39:29 crc kubenswrapper[4815]: E0225 13:39:29.999778 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerName="ceilometer-notification-agent" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.999787 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerName="ceilometer-notification-agent" Feb 25 13:39:29 crc kubenswrapper[4815]: E0225 13:39:29.999801 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d076fbb-4499-456f-82ee-211a8909fc99" containerName="horizon-log" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.999808 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d076fbb-4499-456f-82ee-211a8909fc99" containerName="horizon-log" Feb 25 13:39:29 crc kubenswrapper[4815]: E0225 13:39:29.999826 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerName="ceilometer-central-agent" Feb 25 13:39:29 crc kubenswrapper[4815]: I0225 13:39:29.999833 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerName="ceilometer-central-agent" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.000031 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d076fbb-4499-456f-82ee-211a8909fc99" containerName="horizon-log" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.000057 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d076fbb-4499-456f-82ee-211a8909fc99" containerName="horizon" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.000076 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerName="sg-core" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.000092 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerName="ceilometer-notification-agent" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.000105 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerName="ceilometer-central-agent" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.000121 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="647b5718-0f7b-42e3-aadc-209a60dfc054" containerName="proxy-httpd" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.002131 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.006976 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.009231 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.013750 4815 scope.go:117] "RemoveContainer" containerID="19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.027001 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.044163 4815 scope.go:117] "RemoveContainer" containerID="bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.102268 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-scripts\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.102359 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.102392 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-config-data\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.102426 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-log-httpd\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.102531 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svg2v\" (UniqueName: \"kubernetes.io/projected/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-kube-api-access-svg2v\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.102567 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-run-httpd\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.102603 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.114787 4815 scope.go:117] "RemoveContainer" containerID="7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28" Feb 25 13:39:30 crc kubenswrapper[4815]: E0225 13:39:30.115486 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28\": container with ID starting with 7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28 not found: ID does not exist" containerID="7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.115555 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28"} err="failed to get container status \"7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28\": rpc error: code = NotFound desc = could not find container \"7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28\": container with ID starting with 7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28 not found: ID does not exist" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.115586 4815 scope.go:117] "RemoveContainer" containerID="b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242" Feb 25 13:39:30 crc kubenswrapper[4815]: E0225 13:39:30.116142 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242\": container with ID starting with b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242 not found: ID does not exist" containerID="b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.116181 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242"} err="failed to get container status \"b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242\": rpc error: code = NotFound desc = could not find container \"b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242\": container with ID starting with b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242 not found: ID does not exist" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.116211 4815 scope.go:117] "RemoveContainer" containerID="19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887" Feb 25 13:39:30 crc kubenswrapper[4815]: E0225 13:39:30.116587 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887\": container with ID starting with 19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887 not found: ID does not exist" containerID="19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.116622 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887"} err="failed to get container status \"19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887\": rpc error: code = NotFound desc = could not find container \"19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887\": container with ID starting with 19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887 not found: ID does not exist" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.116642 4815 scope.go:117] "RemoveContainer" containerID="bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19" Feb 25 13:39:30 crc kubenswrapper[4815]: E0225 13:39:30.116913 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19\": container with ID starting with bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19 not found: ID does not exist" containerID="bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.116940 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19"} err="failed to get container status \"bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19\": rpc error: code = NotFound desc = could not find container \"bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19\": container with ID starting with bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19 not found: ID does not exist" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.116957 4815 scope.go:117] "RemoveContainer" containerID="7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.117267 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28"} err="failed to get container status \"7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28\": rpc error: code = NotFound desc = could not find container \"7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28\": container with ID starting with 7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28 not found: ID does not exist" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.117295 4815 scope.go:117] "RemoveContainer" containerID="b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.117617 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242"} err="failed to get container status \"b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242\": rpc error: code = NotFound desc = could not find container \"b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242\": container with ID starting with b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242 not found: ID does not exist" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.117644 4815 scope.go:117] "RemoveContainer" containerID="19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.118227 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887"} err="failed to get container status \"19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887\": rpc error: code = NotFound desc = could not find container \"19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887\": container with ID starting with 19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887 not found: ID does not exist" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.118250 4815 scope.go:117] "RemoveContainer" containerID="bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.118541 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19"} err="failed to get container status \"bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19\": rpc error: code = NotFound desc = could not find container \"bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19\": container with ID starting with bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19 not found: ID does not exist" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.118565 4815 scope.go:117] "RemoveContainer" containerID="7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.119320 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28"} err="failed to get container status \"7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28\": rpc error: code = NotFound desc = could not find container \"7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28\": container with ID starting with 7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28 not found: ID does not exist" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.119341 4815 scope.go:117] "RemoveContainer" containerID="b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.119825 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242"} err="failed to get container status \"b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242\": rpc error: code = NotFound desc = could not find container \"b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242\": container with ID starting with b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242 not found: ID does not exist" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.119846 4815 scope.go:117] "RemoveContainer" containerID="19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.120459 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887"} err="failed to get container status \"19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887\": rpc error: code = NotFound desc = could not find container \"19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887\": container with ID starting with 19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887 not found: ID does not exist" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.120481 4815 scope.go:117] "RemoveContainer" containerID="bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.120780 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19"} err="failed to get container status \"bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19\": rpc error: code = NotFound desc = could not find container \"bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19\": container with ID starting with bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19 not found: ID does not exist" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.120803 4815 scope.go:117] "RemoveContainer" containerID="7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.121605 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28"} err="failed to get container status \"7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28\": rpc error: code = NotFound desc = could not find container \"7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28\": container with ID starting with 7851f0777c2e6b14b437e32542caee955dcd83996f736af52169de007b160f28 not found: ID does not exist" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.121625 4815 scope.go:117] "RemoveContainer" containerID="b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.121815 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242"} err="failed to get container status \"b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242\": rpc error: code = NotFound desc = could not find container \"b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242\": container with ID starting with b3891ae505e23104ed1575c1dfa956030c4c1e1aab23f78ee1c1de5a64144242 not found: ID does not exist" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.121837 4815 scope.go:117] "RemoveContainer" containerID="19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.122009 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887"} err="failed to get container status \"19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887\": rpc error: code = NotFound desc = could not find container \"19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887\": container with ID starting with 19b7a2b0766f2d1a80d29c536a2841e4660ddb48e0d4ac5952c7011240116887 not found: ID does not exist" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.122029 4815 scope.go:117] "RemoveContainer" containerID="bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.122209 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19"} err="failed to get container status \"bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19\": rpc error: code = NotFound desc = could not find container \"bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19\": container with ID starting with bc68b0e271a19ddaec26f0d336b1b878115481ff0cee2c6660400f398145db19 not found: ID does not exist" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.204498 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svg2v\" (UniqueName: \"kubernetes.io/projected/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-kube-api-access-svg2v\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.204558 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-run-httpd\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.204584 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.204661 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-scripts\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.204697 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.204714 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-config-data\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.204735 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-log-httpd\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.205228 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-log-httpd\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.206196 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-run-httpd\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.209125 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.209135 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-scripts\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.210382 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-config-data\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.220272 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.236211 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svg2v\" (UniqueName: \"kubernetes.io/projected/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-kube-api-access-svg2v\") pod \"ceilometer-0\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.321925 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:39:30 crc kubenswrapper[4815]: W0225 13:39:30.761999 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b1fe4db_994d_45e8_8fe6_cf7ae48fceed.slice/crio-a7f05c9da9cae3ff888e5e9487fc09da612cb8fcffb7fb07a041941c0a40f4f7 WatchSource:0}: Error finding container a7f05c9da9cae3ff888e5e9487fc09da612cb8fcffb7fb07a041941c0a40f4f7: Status 404 returned error can't find the container with id a7f05c9da9cae3ff888e5e9487fc09da612cb8fcffb7fb07a041941c0a40f4f7 Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.763725 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.953218 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="647b5718-0f7b-42e3-aadc-209a60dfc054" path="/var/lib/kubelet/pods/647b5718-0f7b-42e3-aadc-209a60dfc054/volumes" Feb 25 13:39:30 crc kubenswrapper[4815]: I0225 13:39:30.954869 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed","Type":"ContainerStarted","Data":"a7f05c9da9cae3ff888e5e9487fc09da612cb8fcffb7fb07a041941c0a40f4f7"} Feb 25 13:39:31 crc kubenswrapper[4815]: I0225 13:39:31.965314 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed","Type":"ContainerStarted","Data":"edd7ba3e3bec4a924a0663ea6db81a4725516914bfc2ea002eef9f39277a222b"} Feb 25 13:39:32 crc kubenswrapper[4815]: I0225 13:39:32.979849 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed","Type":"ContainerStarted","Data":"ebc9650929c8cfee47b3fde90df73dcaa7d8b5303b9040f866f34cf7baad7c11"} Feb 25 13:39:32 crc kubenswrapper[4815]: I0225 13:39:32.980280 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed","Type":"ContainerStarted","Data":"9a826141b8943d53a8ad01a3e1113afb3f2a82d96f38b33fd022f2f6ecb45b61"} Feb 25 13:39:36 crc kubenswrapper[4815]: I0225 13:39:36.014301 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed","Type":"ContainerStarted","Data":"cb93c29ee72beab58cc1c2733f9d6bd896e85d73b56339975285fece1a1964c4"} Feb 25 13:39:36 crc kubenswrapper[4815]: I0225 13:39:36.014819 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 13:39:36 crc kubenswrapper[4815]: I0225 13:39:36.047190 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.49752929 podStartE2EDuration="7.047163812s" podCreationTimestamp="2026-02-25 13:39:29 +0000 UTC" firstStartedPulling="2026-02-25 13:39:30.764370707 +0000 UTC m=+1128.565468761" lastFinishedPulling="2026-02-25 13:39:35.314005229 +0000 UTC m=+1133.115103283" observedRunningTime="2026-02-25 13:39:36.040924396 +0000 UTC m=+1133.842022480" watchObservedRunningTime="2026-02-25 13:39:36.047163812 +0000 UTC m=+1133.848261876" Feb 25 13:39:37 crc kubenswrapper[4815]: I0225 13:39:37.024046 4815 generic.go:334] "Generic (PLEG): container finished" podID="e2580f26-4938-40dc-8a0f-044232a70469" containerID="ce07eb1dae824de5b0f51fe31b9a37f69fba77d6ee95300dc5ae30b254ebf417" exitCode=0 Feb 25 13:39:37 crc kubenswrapper[4815]: I0225 13:39:37.024170 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6t9rm" event={"ID":"e2580f26-4938-40dc-8a0f-044232a70469","Type":"ContainerDied","Data":"ce07eb1dae824de5b0f51fe31b9a37f69fba77d6ee95300dc5ae30b254ebf417"} Feb 25 13:39:38 crc kubenswrapper[4815]: I0225 13:39:38.449957 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6t9rm" Feb 25 13:39:38 crc kubenswrapper[4815]: I0225 13:39:38.566473 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-scripts\") pod \"e2580f26-4938-40dc-8a0f-044232a70469\" (UID: \"e2580f26-4938-40dc-8a0f-044232a70469\") " Feb 25 13:39:38 crc kubenswrapper[4815]: I0225 13:39:38.566565 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vrmx\" (UniqueName: \"kubernetes.io/projected/e2580f26-4938-40dc-8a0f-044232a70469-kube-api-access-7vrmx\") pod \"e2580f26-4938-40dc-8a0f-044232a70469\" (UID: \"e2580f26-4938-40dc-8a0f-044232a70469\") " Feb 25 13:39:38 crc kubenswrapper[4815]: I0225 13:39:38.566736 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-combined-ca-bundle\") pod \"e2580f26-4938-40dc-8a0f-044232a70469\" (UID: \"e2580f26-4938-40dc-8a0f-044232a70469\") " Feb 25 13:39:38 crc kubenswrapper[4815]: I0225 13:39:38.566831 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-config-data\") pod \"e2580f26-4938-40dc-8a0f-044232a70469\" (UID: \"e2580f26-4938-40dc-8a0f-044232a70469\") " Feb 25 13:39:38 crc kubenswrapper[4815]: I0225 13:39:38.572216 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2580f26-4938-40dc-8a0f-044232a70469-kube-api-access-7vrmx" (OuterVolumeSpecName: "kube-api-access-7vrmx") pod "e2580f26-4938-40dc-8a0f-044232a70469" (UID: "e2580f26-4938-40dc-8a0f-044232a70469"). InnerVolumeSpecName "kube-api-access-7vrmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:39:38 crc kubenswrapper[4815]: I0225 13:39:38.573559 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-scripts" (OuterVolumeSpecName: "scripts") pod "e2580f26-4938-40dc-8a0f-044232a70469" (UID: "e2580f26-4938-40dc-8a0f-044232a70469"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:38 crc kubenswrapper[4815]: I0225 13:39:38.595701 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-config-data" (OuterVolumeSpecName: "config-data") pod "e2580f26-4938-40dc-8a0f-044232a70469" (UID: "e2580f26-4938-40dc-8a0f-044232a70469"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:38 crc kubenswrapper[4815]: I0225 13:39:38.605909 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2580f26-4938-40dc-8a0f-044232a70469" (UID: "e2580f26-4938-40dc-8a0f-044232a70469"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:38 crc kubenswrapper[4815]: I0225 13:39:38.669383 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:38 crc kubenswrapper[4815]: I0225 13:39:38.669422 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:38 crc kubenswrapper[4815]: I0225 13:39:38.669434 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vrmx\" (UniqueName: \"kubernetes.io/projected/e2580f26-4938-40dc-8a0f-044232a70469-kube-api-access-7vrmx\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:38 crc kubenswrapper[4815]: I0225 13:39:38.669446 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2580f26-4938-40dc-8a0f-044232a70469-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.041330 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6t9rm" event={"ID":"e2580f26-4938-40dc-8a0f-044232a70469","Type":"ContainerDied","Data":"c0ce3ac48de315f34c630b709d75944a991e1a5ccde0433004e6b765023ba1d5"} Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.041604 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0ce3ac48de315f34c630b709d75944a991e1a5ccde0433004e6b765023ba1d5" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.041377 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6t9rm" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.146663 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 13:39:39 crc kubenswrapper[4815]: E0225 13:39:39.147810 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2580f26-4938-40dc-8a0f-044232a70469" containerName="nova-cell0-conductor-db-sync" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.147939 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2580f26-4938-40dc-8a0f-044232a70469" containerName="nova-cell0-conductor-db-sync" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.148526 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2580f26-4938-40dc-8a0f-044232a70469" containerName="nova-cell0-conductor-db-sync" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.150986 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.163158 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.163406 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-lstls" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.166352 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.281212 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42f9d694-0f9e-4754-86e6-7301c4485eaa-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"42f9d694-0f9e-4754-86e6-7301c4485eaa\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.281349 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42f9d694-0f9e-4754-86e6-7301c4485eaa-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"42f9d694-0f9e-4754-86e6-7301c4485eaa\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.281432 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86bsz\" (UniqueName: \"kubernetes.io/projected/42f9d694-0f9e-4754-86e6-7301c4485eaa-kube-api-access-86bsz\") pod \"nova-cell0-conductor-0\" (UID: \"42f9d694-0f9e-4754-86e6-7301c4485eaa\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.383149 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42f9d694-0f9e-4754-86e6-7301c4485eaa-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"42f9d694-0f9e-4754-86e6-7301c4485eaa\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.383396 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86bsz\" (UniqueName: \"kubernetes.io/projected/42f9d694-0f9e-4754-86e6-7301c4485eaa-kube-api-access-86bsz\") pod \"nova-cell0-conductor-0\" (UID: \"42f9d694-0f9e-4754-86e6-7301c4485eaa\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.383527 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42f9d694-0f9e-4754-86e6-7301c4485eaa-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"42f9d694-0f9e-4754-86e6-7301c4485eaa\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.388548 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42f9d694-0f9e-4754-86e6-7301c4485eaa-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"42f9d694-0f9e-4754-86e6-7301c4485eaa\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.396986 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42f9d694-0f9e-4754-86e6-7301c4485eaa-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"42f9d694-0f9e-4754-86e6-7301c4485eaa\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.403946 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86bsz\" (UniqueName: \"kubernetes.io/projected/42f9d694-0f9e-4754-86e6-7301c4485eaa-kube-api-access-86bsz\") pod \"nova-cell0-conductor-0\" (UID: \"42f9d694-0f9e-4754-86e6-7301c4485eaa\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.474935 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.657349 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 13:39:39 crc kubenswrapper[4815]: W0225 13:39:39.963708 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42f9d694_0f9e_4754_86e6_7301c4485eaa.slice/crio-a15d0b5d5b9285b9d26d843bc7e086925ec85a21ceed15017d36afdc6992460d WatchSource:0}: Error finding container a15d0b5d5b9285b9d26d843bc7e086925ec85a21ceed15017d36afdc6992460d: Status 404 returned error can't find the container with id a15d0b5d5b9285b9d26d843bc7e086925ec85a21ceed15017d36afdc6992460d Feb 25 13:39:39 crc kubenswrapper[4815]: I0225 13:39:39.969313 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 13:39:40 crc kubenswrapper[4815]: I0225 13:39:40.051079 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"42f9d694-0f9e-4754-86e6-7301c4485eaa","Type":"ContainerStarted","Data":"a15d0b5d5b9285b9d26d843bc7e086925ec85a21ceed15017d36afdc6992460d"} Feb 25 13:39:41 crc kubenswrapper[4815]: I0225 13:39:41.063396 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"42f9d694-0f9e-4754-86e6-7301c4485eaa","Type":"ContainerStarted","Data":"c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4"} Feb 25 13:39:41 crc kubenswrapper[4815]: I0225 13:39:41.063942 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 25 13:39:41 crc kubenswrapper[4815]: I0225 13:39:41.063616 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="42f9d694-0f9e-4754-86e6-7301c4485eaa" containerName="nova-cell0-conductor-conductor" containerID="cri-o://c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" gracePeriod=30 Feb 25 13:39:41 crc kubenswrapper[4815]: I0225 13:39:41.101607 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.10157786 podStartE2EDuration="2.10157786s" podCreationTimestamp="2026-02-25 13:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:39:41.087053116 +0000 UTC m=+1138.888151200" watchObservedRunningTime="2026-02-25 13:39:41.10157786 +0000 UTC m=+1138.902675944" Feb 25 13:39:41 crc kubenswrapper[4815]: I0225 13:39:41.427262 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:41 crc kubenswrapper[4815]: I0225 13:39:41.427865 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerName="ceilometer-central-agent" containerID="cri-o://edd7ba3e3bec4a924a0663ea6db81a4725516914bfc2ea002eef9f39277a222b" gracePeriod=30 Feb 25 13:39:41 crc kubenswrapper[4815]: I0225 13:39:41.427894 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerName="proxy-httpd" containerID="cri-o://cb93c29ee72beab58cc1c2733f9d6bd896e85d73b56339975285fece1a1964c4" gracePeriod=30 Feb 25 13:39:41 crc kubenswrapper[4815]: I0225 13:39:41.427926 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerName="sg-core" containerID="cri-o://ebc9650929c8cfee47b3fde90df73dcaa7d8b5303b9040f866f34cf7baad7c11" gracePeriod=30 Feb 25 13:39:41 crc kubenswrapper[4815]: I0225 13:39:41.427973 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerName="ceilometer-notification-agent" containerID="cri-o://9a826141b8943d53a8ad01a3e1113afb3f2a82d96f38b33fd022f2f6ecb45b61" gracePeriod=30 Feb 25 13:39:42 crc kubenswrapper[4815]: E0225 13:39:42.023962 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b1fe4db_994d_45e8_8fe6_cf7ae48fceed.slice/crio-edd7ba3e3bec4a924a0663ea6db81a4725516914bfc2ea002eef9f39277a222b.scope\": RecentStats: unable to find data in memory cache]" Feb 25 13:39:42 crc kubenswrapper[4815]: I0225 13:39:42.073472 4815 generic.go:334] "Generic (PLEG): container finished" podID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerID="cb93c29ee72beab58cc1c2733f9d6bd896e85d73b56339975285fece1a1964c4" exitCode=0 Feb 25 13:39:42 crc kubenswrapper[4815]: I0225 13:39:42.073542 4815 generic.go:334] "Generic (PLEG): container finished" podID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerID="ebc9650929c8cfee47b3fde90df73dcaa7d8b5303b9040f866f34cf7baad7c11" exitCode=2 Feb 25 13:39:42 crc kubenswrapper[4815]: I0225 13:39:42.073553 4815 generic.go:334] "Generic (PLEG): container finished" podID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerID="edd7ba3e3bec4a924a0663ea6db81a4725516914bfc2ea002eef9f39277a222b" exitCode=0 Feb 25 13:39:42 crc kubenswrapper[4815]: I0225 13:39:42.073543 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed","Type":"ContainerDied","Data":"cb93c29ee72beab58cc1c2733f9d6bd896e85d73b56339975285fece1a1964c4"} Feb 25 13:39:42 crc kubenswrapper[4815]: I0225 13:39:42.073598 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed","Type":"ContainerDied","Data":"ebc9650929c8cfee47b3fde90df73dcaa7d8b5303b9040f866f34cf7baad7c11"} Feb 25 13:39:42 crc kubenswrapper[4815]: I0225 13:39:42.073626 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed","Type":"ContainerDied","Data":"edd7ba3e3bec4a924a0663ea6db81a4725516914bfc2ea002eef9f39277a222b"} Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.085677 4815 generic.go:334] "Generic (PLEG): container finished" podID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerID="9a826141b8943d53a8ad01a3e1113afb3f2a82d96f38b33fd022f2f6ecb45b61" exitCode=0 Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.085757 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed","Type":"ContainerDied","Data":"9a826141b8943d53a8ad01a3e1113afb3f2a82d96f38b33fd022f2f6ecb45b61"} Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.211202 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.366301 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-scripts\") pod \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.366478 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-config-data\") pod \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.366534 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-run-httpd\") pod \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.366555 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-combined-ca-bundle\") pod \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.366600 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-log-httpd\") pod \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.366646 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svg2v\" (UniqueName: \"kubernetes.io/projected/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-kube-api-access-svg2v\") pod \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.366732 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-sg-core-conf-yaml\") pod \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\" (UID: \"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed\") " Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.367296 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" (UID: "0b1fe4db-994d-45e8-8fe6-cf7ae48fceed"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.367887 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" (UID: "0b1fe4db-994d-45e8-8fe6-cf7ae48fceed"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.376697 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-scripts" (OuterVolumeSpecName: "scripts") pod "0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" (UID: "0b1fe4db-994d-45e8-8fe6-cf7ae48fceed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.376780 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-kube-api-access-svg2v" (OuterVolumeSpecName: "kube-api-access-svg2v") pod "0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" (UID: "0b1fe4db-994d-45e8-8fe6-cf7ae48fceed"). InnerVolumeSpecName "kube-api-access-svg2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.397987 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" (UID: "0b1fe4db-994d-45e8-8fe6-cf7ae48fceed"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.456317 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" (UID: "0b1fe4db-994d-45e8-8fe6-cf7ae48fceed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.467933 4815 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.467970 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svg2v\" (UniqueName: \"kubernetes.io/projected/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-kube-api-access-svg2v\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.467983 4815 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.467993 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.468001 4815 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.468009 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.490174 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-config-data" (OuterVolumeSpecName: "config-data") pod "0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" (UID: "0b1fe4db-994d-45e8-8fe6-cf7ae48fceed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:39:43 crc kubenswrapper[4815]: I0225 13:39:43.569990 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.106861 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0b1fe4db-994d-45e8-8fe6-cf7ae48fceed","Type":"ContainerDied","Data":"a7f05c9da9cae3ff888e5e9487fc09da612cb8fcffb7fb07a041941c0a40f4f7"} Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.106928 4815 scope.go:117] "RemoveContainer" containerID="cb93c29ee72beab58cc1c2733f9d6bd896e85d73b56339975285fece1a1964c4" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.106970 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.157355 4815 scope.go:117] "RemoveContainer" containerID="ebc9650929c8cfee47b3fde90df73dcaa7d8b5303b9040f866f34cf7baad7c11" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.183473 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.201053 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.201123 4815 scope.go:117] "RemoveContainer" containerID="9a826141b8943d53a8ad01a3e1113afb3f2a82d96f38b33fd022f2f6ecb45b61" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.217555 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:44 crc kubenswrapper[4815]: E0225 13:39:44.218130 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerName="ceilometer-notification-agent" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.218161 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerName="ceilometer-notification-agent" Feb 25 13:39:44 crc kubenswrapper[4815]: E0225 13:39:44.218187 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerName="ceilometer-central-agent" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.218197 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerName="ceilometer-central-agent" Feb 25 13:39:44 crc kubenswrapper[4815]: E0225 13:39:44.218208 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerName="proxy-httpd" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.218217 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerName="proxy-httpd" Feb 25 13:39:44 crc kubenswrapper[4815]: E0225 13:39:44.218232 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerName="sg-core" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.218240 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerName="sg-core" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.218485 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerName="ceilometer-central-agent" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.218501 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerName="proxy-httpd" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.218550 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerName="sg-core" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.218573 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" containerName="ceilometer-notification-agent" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.221056 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.228034 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.228435 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.237960 4815 scope.go:117] "RemoveContainer" containerID="edd7ba3e3bec4a924a0663ea6db81a4725516914bfc2ea002eef9f39277a222b" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.245157 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.286612 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.286700 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.286738 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/922cea47-8f1c-4f9d-bbd4-f7a82c364356-log-httpd\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.286861 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/922cea47-8f1c-4f9d-bbd4-f7a82c364356-run-httpd\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.286907 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-scripts\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.287002 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-config-data\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.287032 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn2vt\" (UniqueName: \"kubernetes.io/projected/922cea47-8f1c-4f9d-bbd4-f7a82c364356-kube-api-access-jn2vt\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.388566 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.388714 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.388776 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/922cea47-8f1c-4f9d-bbd4-f7a82c364356-log-httpd\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.388856 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/922cea47-8f1c-4f9d-bbd4-f7a82c364356-run-httpd\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.388899 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-scripts\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.389029 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-config-data\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.389063 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn2vt\" (UniqueName: \"kubernetes.io/projected/922cea47-8f1c-4f9d-bbd4-f7a82c364356-kube-api-access-jn2vt\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.390909 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/922cea47-8f1c-4f9d-bbd4-f7a82c364356-run-httpd\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.391811 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/922cea47-8f1c-4f9d-bbd4-f7a82c364356-log-httpd\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.395685 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-scripts\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.398321 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.398498 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-config-data\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.398746 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.406805 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn2vt\" (UniqueName: \"kubernetes.io/projected/922cea47-8f1c-4f9d-bbd4-f7a82c364356-kube-api-access-jn2vt\") pod \"ceilometer-0\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.550450 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:39:44 crc kubenswrapper[4815]: I0225 13:39:44.951899 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b1fe4db-994d-45e8-8fe6-cf7ae48fceed" path="/var/lib/kubelet/pods/0b1fe4db-994d-45e8-8fe6-cf7ae48fceed/volumes" Feb 25 13:39:45 crc kubenswrapper[4815]: I0225 13:39:45.094213 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:39:45 crc kubenswrapper[4815]: I0225 13:39:45.127777 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"922cea47-8f1c-4f9d-bbd4-f7a82c364356","Type":"ContainerStarted","Data":"0f2aefb641682236d9d5f46024d7cfd956ac7ccea175a9213049a8ed5e79e9d5"} Feb 25 13:39:46 crc kubenswrapper[4815]: I0225 13:39:46.141355 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"922cea47-8f1c-4f9d-bbd4-f7a82c364356","Type":"ContainerStarted","Data":"9400070e3a927941ad9839f3539cd29755bf156d96c7eb6767afd33b839509ad"} Feb 25 13:39:47 crc kubenswrapper[4815]: I0225 13:39:47.160747 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"922cea47-8f1c-4f9d-bbd4-f7a82c364356","Type":"ContainerStarted","Data":"5df069ce02edb3750aaa11eb2126eb7af892e52967faa9f70cdd8a79bf0a1dfe"} Feb 25 13:39:48 crc kubenswrapper[4815]: I0225 13:39:48.192239 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"922cea47-8f1c-4f9d-bbd4-f7a82c364356","Type":"ContainerStarted","Data":"379123921ac18dbae4ad9a5710dbbdfda9c066f6bde614a4ab7bd51a5ae51958"} Feb 25 13:39:49 crc kubenswrapper[4815]: I0225 13:39:49.207470 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"922cea47-8f1c-4f9d-bbd4-f7a82c364356","Type":"ContainerStarted","Data":"e7af682dbfbb6a00b57a0296b16bc311e4ee020e018576c68f0a1cc9b32f4c9b"} Feb 25 13:39:49 crc kubenswrapper[4815]: I0225 13:39:49.208348 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 13:39:49 crc kubenswrapper[4815]: I0225 13:39:49.238592 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.607428525 podStartE2EDuration="5.238573399s" podCreationTimestamp="2026-02-25 13:39:44 +0000 UTC" firstStartedPulling="2026-02-25 13:39:45.115142026 +0000 UTC m=+1142.916240080" lastFinishedPulling="2026-02-25 13:39:48.74628689 +0000 UTC m=+1146.547384954" observedRunningTime="2026-02-25 13:39:49.233193356 +0000 UTC m=+1147.034291420" watchObservedRunningTime="2026-02-25 13:39:49.238573399 +0000 UTC m=+1147.039671443" Feb 25 13:39:49 crc kubenswrapper[4815]: E0225 13:39:49.478596 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 13:39:49 crc kubenswrapper[4815]: E0225 13:39:49.528721 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 13:39:49 crc kubenswrapper[4815]: E0225 13:39:49.534056 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 13:39:49 crc kubenswrapper[4815]: E0225 13:39:49.534290 4815 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="42f9d694-0f9e-4754-86e6-7301c4485eaa" containerName="nova-cell0-conductor-conductor" Feb 25 13:39:51 crc kubenswrapper[4815]: I0225 13:39:51.342742 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:39:51 crc kubenswrapper[4815]: I0225 13:39:51.343378 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:39:54 crc kubenswrapper[4815]: E0225 13:39:54.477963 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 13:39:54 crc kubenswrapper[4815]: E0225 13:39:54.480914 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 13:39:54 crc kubenswrapper[4815]: E0225 13:39:54.483505 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 13:39:54 crc kubenswrapper[4815]: E0225 13:39:54.483575 4815 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="42f9d694-0f9e-4754-86e6-7301c4485eaa" containerName="nova-cell0-conductor-conductor" Feb 25 13:39:59 crc kubenswrapper[4815]: E0225 13:39:59.477625 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 13:39:59 crc kubenswrapper[4815]: E0225 13:39:59.480504 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 13:39:59 crc kubenswrapper[4815]: E0225 13:39:59.483055 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 13:39:59 crc kubenswrapper[4815]: E0225 13:39:59.483137 4815 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="42f9d694-0f9e-4754-86e6-7301c4485eaa" containerName="nova-cell0-conductor-conductor" Feb 25 13:40:00 crc kubenswrapper[4815]: I0225 13:40:00.186181 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533780-vbrwb"] Feb 25 13:40:00 crc kubenswrapper[4815]: I0225 13:40:00.187690 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533780-vbrwb" Feb 25 13:40:00 crc kubenswrapper[4815]: I0225 13:40:00.190953 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:40:00 crc kubenswrapper[4815]: I0225 13:40:00.191272 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:40:00 crc kubenswrapper[4815]: I0225 13:40:00.191485 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:40:00 crc kubenswrapper[4815]: I0225 13:40:00.209389 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533780-vbrwb"] Feb 25 13:40:00 crc kubenswrapper[4815]: I0225 13:40:00.240267 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2bsg\" (UniqueName: \"kubernetes.io/projected/a63b31e2-6a8c-4664-a47c-30f2a731c8a4-kube-api-access-g2bsg\") pod \"auto-csr-approver-29533780-vbrwb\" (UID: \"a63b31e2-6a8c-4664-a47c-30f2a731c8a4\") " pod="openshift-infra/auto-csr-approver-29533780-vbrwb" Feb 25 13:40:00 crc kubenswrapper[4815]: I0225 13:40:00.341496 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2bsg\" (UniqueName: \"kubernetes.io/projected/a63b31e2-6a8c-4664-a47c-30f2a731c8a4-kube-api-access-g2bsg\") pod \"auto-csr-approver-29533780-vbrwb\" (UID: \"a63b31e2-6a8c-4664-a47c-30f2a731c8a4\") " pod="openshift-infra/auto-csr-approver-29533780-vbrwb" Feb 25 13:40:00 crc kubenswrapper[4815]: I0225 13:40:00.363527 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2bsg\" (UniqueName: \"kubernetes.io/projected/a63b31e2-6a8c-4664-a47c-30f2a731c8a4-kube-api-access-g2bsg\") pod \"auto-csr-approver-29533780-vbrwb\" (UID: \"a63b31e2-6a8c-4664-a47c-30f2a731c8a4\") " pod="openshift-infra/auto-csr-approver-29533780-vbrwb" Feb 25 13:40:00 crc kubenswrapper[4815]: I0225 13:40:00.517478 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533780-vbrwb" Feb 25 13:40:00 crc kubenswrapper[4815]: I0225 13:40:00.997905 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533780-vbrwb"] Feb 25 13:40:01 crc kubenswrapper[4815]: I0225 13:40:01.344975 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533780-vbrwb" event={"ID":"a63b31e2-6a8c-4664-a47c-30f2a731c8a4","Type":"ContainerStarted","Data":"296bb9413907fa6113b03fcf5c9816fcf78197dbbc765964a229260347d59f29"} Feb 25 13:40:03 crc kubenswrapper[4815]: I0225 13:40:03.370613 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533780-vbrwb" event={"ID":"a63b31e2-6a8c-4664-a47c-30f2a731c8a4","Type":"ContainerStarted","Data":"18dd34d53e550f261ca45eb8f3d6ec974c09281859b2802e967d8144d1bfbda6"} Feb 25 13:40:03 crc kubenswrapper[4815]: I0225 13:40:03.401031 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533780-vbrwb" podStartSLOduration=1.463977006 podStartE2EDuration="3.401004379s" podCreationTimestamp="2026-02-25 13:40:00 +0000 UTC" firstStartedPulling="2026-02-25 13:40:01.006438535 +0000 UTC m=+1158.807536599" lastFinishedPulling="2026-02-25 13:40:02.943465888 +0000 UTC m=+1160.744563972" observedRunningTime="2026-02-25 13:40:03.387612607 +0000 UTC m=+1161.188710661" watchObservedRunningTime="2026-02-25 13:40:03.401004379 +0000 UTC m=+1161.202102473" Feb 25 13:40:04 crc kubenswrapper[4815]: I0225 13:40:04.390954 4815 generic.go:334] "Generic (PLEG): container finished" podID="a63b31e2-6a8c-4664-a47c-30f2a731c8a4" containerID="18dd34d53e550f261ca45eb8f3d6ec974c09281859b2802e967d8144d1bfbda6" exitCode=0 Feb 25 13:40:04 crc kubenswrapper[4815]: I0225 13:40:04.391058 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533780-vbrwb" event={"ID":"a63b31e2-6a8c-4664-a47c-30f2a731c8a4","Type":"ContainerDied","Data":"18dd34d53e550f261ca45eb8f3d6ec974c09281859b2802e967d8144d1bfbda6"} Feb 25 13:40:04 crc kubenswrapper[4815]: E0225 13:40:04.476994 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 13:40:04 crc kubenswrapper[4815]: E0225 13:40:04.478455 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 13:40:04 crc kubenswrapper[4815]: E0225 13:40:04.479961 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 13:40:04 crc kubenswrapper[4815]: E0225 13:40:04.480030 4815 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="42f9d694-0f9e-4754-86e6-7301c4485eaa" containerName="nova-cell0-conductor-conductor" Feb 25 13:40:05 crc kubenswrapper[4815]: I0225 13:40:05.833716 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533780-vbrwb" Feb 25 13:40:05 crc kubenswrapper[4815]: I0225 13:40:05.857349 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2bsg\" (UniqueName: \"kubernetes.io/projected/a63b31e2-6a8c-4664-a47c-30f2a731c8a4-kube-api-access-g2bsg\") pod \"a63b31e2-6a8c-4664-a47c-30f2a731c8a4\" (UID: \"a63b31e2-6a8c-4664-a47c-30f2a731c8a4\") " Feb 25 13:40:05 crc kubenswrapper[4815]: I0225 13:40:05.868936 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a63b31e2-6a8c-4664-a47c-30f2a731c8a4-kube-api-access-g2bsg" (OuterVolumeSpecName: "kube-api-access-g2bsg") pod "a63b31e2-6a8c-4664-a47c-30f2a731c8a4" (UID: "a63b31e2-6a8c-4664-a47c-30f2a731c8a4"). InnerVolumeSpecName "kube-api-access-g2bsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:40:05 crc kubenswrapper[4815]: I0225 13:40:05.959959 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2bsg\" (UniqueName: \"kubernetes.io/projected/a63b31e2-6a8c-4664-a47c-30f2a731c8a4-kube-api-access-g2bsg\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:06 crc kubenswrapper[4815]: I0225 13:40:06.049763 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533774-tm57d"] Feb 25 13:40:06 crc kubenswrapper[4815]: I0225 13:40:06.061566 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533774-tm57d"] Feb 25 13:40:06 crc kubenswrapper[4815]: I0225 13:40:06.420415 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533780-vbrwb" event={"ID":"a63b31e2-6a8c-4664-a47c-30f2a731c8a4","Type":"ContainerDied","Data":"296bb9413907fa6113b03fcf5c9816fcf78197dbbc765964a229260347d59f29"} Feb 25 13:40:06 crc kubenswrapper[4815]: I0225 13:40:06.420706 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="296bb9413907fa6113b03fcf5c9816fcf78197dbbc765964a229260347d59f29" Feb 25 13:40:06 crc kubenswrapper[4815]: I0225 13:40:06.420498 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533780-vbrwb" Feb 25 13:40:06 crc kubenswrapper[4815]: I0225 13:40:06.955362 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a83fc2e6-4dca-419c-ac69-1eb0f160ba73" path="/var/lib/kubelet/pods/a83fc2e6-4dca-419c-ac69-1eb0f160ba73/volumes" Feb 25 13:40:09 crc kubenswrapper[4815]: E0225 13:40:09.478670 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 13:40:09 crc kubenswrapper[4815]: E0225 13:40:09.480779 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 13:40:09 crc kubenswrapper[4815]: E0225 13:40:09.482447 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 13:40:09 crc kubenswrapper[4815]: E0225 13:40:09.482482 4815 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="42f9d694-0f9e-4754-86e6-7301c4485eaa" containerName="nova-cell0-conductor-conductor" Feb 25 13:40:11 crc kubenswrapper[4815]: I0225 13:40:11.480734 4815 generic.go:334] "Generic (PLEG): container finished" podID="42f9d694-0f9e-4754-86e6-7301c4485eaa" containerID="c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" exitCode=137 Feb 25 13:40:11 crc kubenswrapper[4815]: I0225 13:40:11.480814 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"42f9d694-0f9e-4754-86e6-7301c4485eaa","Type":"ContainerDied","Data":"c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4"} Feb 25 13:40:11 crc kubenswrapper[4815]: I0225 13:40:11.481709 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"42f9d694-0f9e-4754-86e6-7301c4485eaa","Type":"ContainerDied","Data":"a15d0b5d5b9285b9d26d843bc7e086925ec85a21ceed15017d36afdc6992460d"} Feb 25 13:40:11 crc kubenswrapper[4815]: I0225 13:40:11.481741 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a15d0b5d5b9285b9d26d843bc7e086925ec85a21ceed15017d36afdc6992460d" Feb 25 13:40:11 crc kubenswrapper[4815]: I0225 13:40:11.536815 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 25 13:40:11 crc kubenswrapper[4815]: I0225 13:40:11.696943 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86bsz\" (UniqueName: \"kubernetes.io/projected/42f9d694-0f9e-4754-86e6-7301c4485eaa-kube-api-access-86bsz\") pod \"42f9d694-0f9e-4754-86e6-7301c4485eaa\" (UID: \"42f9d694-0f9e-4754-86e6-7301c4485eaa\") " Feb 25 13:40:11 crc kubenswrapper[4815]: I0225 13:40:11.697036 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42f9d694-0f9e-4754-86e6-7301c4485eaa-config-data\") pod \"42f9d694-0f9e-4754-86e6-7301c4485eaa\" (UID: \"42f9d694-0f9e-4754-86e6-7301c4485eaa\") " Feb 25 13:40:11 crc kubenswrapper[4815]: I0225 13:40:11.697088 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42f9d694-0f9e-4754-86e6-7301c4485eaa-combined-ca-bundle\") pod \"42f9d694-0f9e-4754-86e6-7301c4485eaa\" (UID: \"42f9d694-0f9e-4754-86e6-7301c4485eaa\") " Feb 25 13:40:11 crc kubenswrapper[4815]: I0225 13:40:11.703572 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42f9d694-0f9e-4754-86e6-7301c4485eaa-kube-api-access-86bsz" (OuterVolumeSpecName: "kube-api-access-86bsz") pod "42f9d694-0f9e-4754-86e6-7301c4485eaa" (UID: "42f9d694-0f9e-4754-86e6-7301c4485eaa"). InnerVolumeSpecName "kube-api-access-86bsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:40:11 crc kubenswrapper[4815]: I0225 13:40:11.745057 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42f9d694-0f9e-4754-86e6-7301c4485eaa-config-data" (OuterVolumeSpecName: "config-data") pod "42f9d694-0f9e-4754-86e6-7301c4485eaa" (UID: "42f9d694-0f9e-4754-86e6-7301c4485eaa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:11 crc kubenswrapper[4815]: I0225 13:40:11.750706 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42f9d694-0f9e-4754-86e6-7301c4485eaa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42f9d694-0f9e-4754-86e6-7301c4485eaa" (UID: "42f9d694-0f9e-4754-86e6-7301c4485eaa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:11 crc kubenswrapper[4815]: I0225 13:40:11.799747 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86bsz\" (UniqueName: \"kubernetes.io/projected/42f9d694-0f9e-4754-86e6-7301c4485eaa-kube-api-access-86bsz\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:11 crc kubenswrapper[4815]: I0225 13:40:11.799785 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42f9d694-0f9e-4754-86e6-7301c4485eaa-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:11 crc kubenswrapper[4815]: I0225 13:40:11.799803 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42f9d694-0f9e-4754-86e6-7301c4485eaa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.491566 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.531941 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.541537 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.607259 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 13:40:12 crc kubenswrapper[4815]: E0225 13:40:12.608467 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42f9d694-0f9e-4754-86e6-7301c4485eaa" containerName="nova-cell0-conductor-conductor" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.608495 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="42f9d694-0f9e-4754-86e6-7301c4485eaa" containerName="nova-cell0-conductor-conductor" Feb 25 13:40:12 crc kubenswrapper[4815]: E0225 13:40:12.608589 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a63b31e2-6a8c-4664-a47c-30f2a731c8a4" containerName="oc" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.608601 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a63b31e2-6a8c-4664-a47c-30f2a731c8a4" containerName="oc" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.609050 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="a63b31e2-6a8c-4664-a47c-30f2a731c8a4" containerName="oc" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.609105 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="42f9d694-0f9e-4754-86e6-7301c4485eaa" containerName="nova-cell0-conductor-conductor" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.610332 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.613087 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-lstls" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.613565 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.617133 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c99a6f77-ea1d-4cb6-848b-447c2ac666c4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c99a6f77-ea1d-4cb6-848b-447c2ac666c4\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.617269 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlkv4\" (UniqueName: \"kubernetes.io/projected/c99a6f77-ea1d-4cb6-848b-447c2ac666c4-kube-api-access-mlkv4\") pod \"nova-cell0-conductor-0\" (UID: \"c99a6f77-ea1d-4cb6-848b-447c2ac666c4\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.618147 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c99a6f77-ea1d-4cb6-848b-447c2ac666c4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c99a6f77-ea1d-4cb6-848b-447c2ac666c4\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.627251 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.719999 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c99a6f77-ea1d-4cb6-848b-447c2ac666c4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c99a6f77-ea1d-4cb6-848b-447c2ac666c4\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.720208 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c99a6f77-ea1d-4cb6-848b-447c2ac666c4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c99a6f77-ea1d-4cb6-848b-447c2ac666c4\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.720389 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlkv4\" (UniqueName: \"kubernetes.io/projected/c99a6f77-ea1d-4cb6-848b-447c2ac666c4-kube-api-access-mlkv4\") pod \"nova-cell0-conductor-0\" (UID: \"c99a6f77-ea1d-4cb6-848b-447c2ac666c4\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.725719 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c99a6f77-ea1d-4cb6-848b-447c2ac666c4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c99a6f77-ea1d-4cb6-848b-447c2ac666c4\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.726893 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c99a6f77-ea1d-4cb6-848b-447c2ac666c4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c99a6f77-ea1d-4cb6-848b-447c2ac666c4\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.738062 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlkv4\" (UniqueName: \"kubernetes.io/projected/c99a6f77-ea1d-4cb6-848b-447c2ac666c4-kube-api-access-mlkv4\") pod \"nova-cell0-conductor-0\" (UID: \"c99a6f77-ea1d-4cb6-848b-447c2ac666c4\") " pod="openstack/nova-cell0-conductor-0" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.937395 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 25 13:40:12 crc kubenswrapper[4815]: I0225 13:40:12.949149 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42f9d694-0f9e-4754-86e6-7301c4485eaa" path="/var/lib/kubelet/pods/42f9d694-0f9e-4754-86e6-7301c4485eaa/volumes" Feb 25 13:40:13 crc kubenswrapper[4815]: I0225 13:40:13.452168 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 13:40:13 crc kubenswrapper[4815]: W0225 13:40:13.457304 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc99a6f77_ea1d_4cb6_848b_447c2ac666c4.slice/crio-9f867c43daf17c9381341157dcdefd4702468657181e5ef4af3e6bf52b20f57e WatchSource:0}: Error finding container 9f867c43daf17c9381341157dcdefd4702468657181e5ef4af3e6bf52b20f57e: Status 404 returned error can't find the container with id 9f867c43daf17c9381341157dcdefd4702468657181e5ef4af3e6bf52b20f57e Feb 25 13:40:13 crc kubenswrapper[4815]: I0225 13:40:13.503501 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c99a6f77-ea1d-4cb6-848b-447c2ac666c4","Type":"ContainerStarted","Data":"9f867c43daf17c9381341157dcdefd4702468657181e5ef4af3e6bf52b20f57e"} Feb 25 13:40:14 crc kubenswrapper[4815]: I0225 13:40:14.518840 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c99a6f77-ea1d-4cb6-848b-447c2ac666c4","Type":"ContainerStarted","Data":"997d78848e29ebdfa5a9a21c05b542b3e80aed48d23fda31408c0876636ad1f5"} Feb 25 13:40:14 crc kubenswrapper[4815]: I0225 13:40:14.519355 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 25 13:40:14 crc kubenswrapper[4815]: I0225 13:40:14.551919 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.551892766 podStartE2EDuration="2.551892766s" podCreationTimestamp="2026-02-25 13:40:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:40:14.534698005 +0000 UTC m=+1172.335796119" watchObservedRunningTime="2026-02-25 13:40:14.551892766 +0000 UTC m=+1172.352990850" Feb 25 13:40:14 crc kubenswrapper[4815]: I0225 13:40:14.556502 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 25 13:40:18 crc kubenswrapper[4815]: I0225 13:40:18.431418 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 13:40:18 crc kubenswrapper[4815]: I0225 13:40:18.432136 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312" containerName="kube-state-metrics" containerID="cri-o://e7e1d82b05c4847e358e488d560ac8a8df154f0e40cfd12e322ad254ac268410" gracePeriod=30 Feb 25 13:40:18 crc kubenswrapper[4815]: I0225 13:40:18.966046 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.055937 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2phkz\" (UniqueName: \"kubernetes.io/projected/9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312-kube-api-access-2phkz\") pod \"9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312\" (UID: \"9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312\") " Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.062534 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312-kube-api-access-2phkz" (OuterVolumeSpecName: "kube-api-access-2phkz") pod "9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312" (UID: "9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312"). InnerVolumeSpecName "kube-api-access-2phkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.158267 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2phkz\" (UniqueName: \"kubernetes.io/projected/9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312-kube-api-access-2phkz\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.564774 4815 generic.go:334] "Generic (PLEG): container finished" podID="9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312" containerID="e7e1d82b05c4847e358e488d560ac8a8df154f0e40cfd12e322ad254ac268410" exitCode=2 Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.564813 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312","Type":"ContainerDied","Data":"e7e1d82b05c4847e358e488d560ac8a8df154f0e40cfd12e322ad254ac268410"} Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.564839 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312","Type":"ContainerDied","Data":"05da5e98867f3a1239d0b8bca687c4a5e4cebc0a0f2e56aac9e1f9c81e25c716"} Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.564854 4815 scope.go:117] "RemoveContainer" containerID="e7e1d82b05c4847e358e488d560ac8a8df154f0e40cfd12e322ad254ac268410" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.566077 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.597612 4815 scope.go:117] "RemoveContainer" containerID="e7e1d82b05c4847e358e488d560ac8a8df154f0e40cfd12e322ad254ac268410" Feb 25 13:40:19 crc kubenswrapper[4815]: E0225 13:40:19.600692 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7e1d82b05c4847e358e488d560ac8a8df154f0e40cfd12e322ad254ac268410\": container with ID starting with e7e1d82b05c4847e358e488d560ac8a8df154f0e40cfd12e322ad254ac268410 not found: ID does not exist" containerID="e7e1d82b05c4847e358e488d560ac8a8df154f0e40cfd12e322ad254ac268410" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.600745 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7e1d82b05c4847e358e488d560ac8a8df154f0e40cfd12e322ad254ac268410"} err="failed to get container status \"e7e1d82b05c4847e358e488d560ac8a8df154f0e40cfd12e322ad254ac268410\": rpc error: code = NotFound desc = could not find container \"e7e1d82b05c4847e358e488d560ac8a8df154f0e40cfd12e322ad254ac268410\": container with ID starting with e7e1d82b05c4847e358e488d560ac8a8df154f0e40cfd12e322ad254ac268410 not found: ID does not exist" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.622551 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.635985 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.646900 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 13:40:19 crc kubenswrapper[4815]: E0225 13:40:19.647615 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312" containerName="kube-state-metrics" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.647747 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312" containerName="kube-state-metrics" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.648049 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312" containerName="kube-state-metrics" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.648889 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.654715 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.656211 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.656558 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.770564 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1407af2f-82a3-4ccb-9667-03d441ff34dc-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1407af2f-82a3-4ccb-9667-03d441ff34dc\") " pod="openstack/kube-state-metrics-0" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.770969 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1407af2f-82a3-4ccb-9667-03d441ff34dc-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1407af2f-82a3-4ccb-9667-03d441ff34dc\") " pod="openstack/kube-state-metrics-0" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.771145 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ntw8\" (UniqueName: \"kubernetes.io/projected/1407af2f-82a3-4ccb-9667-03d441ff34dc-kube-api-access-8ntw8\") pod \"kube-state-metrics-0\" (UID: \"1407af2f-82a3-4ccb-9667-03d441ff34dc\") " pod="openstack/kube-state-metrics-0" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.771208 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1407af2f-82a3-4ccb-9667-03d441ff34dc-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1407af2f-82a3-4ccb-9667-03d441ff34dc\") " pod="openstack/kube-state-metrics-0" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.873204 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ntw8\" (UniqueName: \"kubernetes.io/projected/1407af2f-82a3-4ccb-9667-03d441ff34dc-kube-api-access-8ntw8\") pod \"kube-state-metrics-0\" (UID: \"1407af2f-82a3-4ccb-9667-03d441ff34dc\") " pod="openstack/kube-state-metrics-0" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.873254 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1407af2f-82a3-4ccb-9667-03d441ff34dc-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1407af2f-82a3-4ccb-9667-03d441ff34dc\") " pod="openstack/kube-state-metrics-0" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.873304 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1407af2f-82a3-4ccb-9667-03d441ff34dc-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1407af2f-82a3-4ccb-9667-03d441ff34dc\") " pod="openstack/kube-state-metrics-0" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.873389 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1407af2f-82a3-4ccb-9667-03d441ff34dc-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1407af2f-82a3-4ccb-9667-03d441ff34dc\") " pod="openstack/kube-state-metrics-0" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.877838 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1407af2f-82a3-4ccb-9667-03d441ff34dc-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1407af2f-82a3-4ccb-9667-03d441ff34dc\") " pod="openstack/kube-state-metrics-0" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.878187 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1407af2f-82a3-4ccb-9667-03d441ff34dc-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1407af2f-82a3-4ccb-9667-03d441ff34dc\") " pod="openstack/kube-state-metrics-0" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.880194 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1407af2f-82a3-4ccb-9667-03d441ff34dc-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1407af2f-82a3-4ccb-9667-03d441ff34dc\") " pod="openstack/kube-state-metrics-0" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.891493 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ntw8\" (UniqueName: \"kubernetes.io/projected/1407af2f-82a3-4ccb-9667-03d441ff34dc-kube-api-access-8ntw8\") pod \"kube-state-metrics-0\" (UID: \"1407af2f-82a3-4ccb-9667-03d441ff34dc\") " pod="openstack/kube-state-metrics-0" Feb 25 13:40:19 crc kubenswrapper[4815]: I0225 13:40:19.969459 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 13:40:20 crc kubenswrapper[4815]: I0225 13:40:20.099556 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:40:20 crc kubenswrapper[4815]: I0225 13:40:20.100175 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerName="ceilometer-central-agent" containerID="cri-o://9400070e3a927941ad9839f3539cd29755bf156d96c7eb6767afd33b839509ad" gracePeriod=30 Feb 25 13:40:20 crc kubenswrapper[4815]: I0225 13:40:20.100557 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerName="proxy-httpd" containerID="cri-o://e7af682dbfbb6a00b57a0296b16bc311e4ee020e018576c68f0a1cc9b32f4c9b" gracePeriod=30 Feb 25 13:40:20 crc kubenswrapper[4815]: I0225 13:40:20.101134 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerName="sg-core" containerID="cri-o://379123921ac18dbae4ad9a5710dbbdfda9c066f6bde614a4ab7bd51a5ae51958" gracePeriod=30 Feb 25 13:40:20 crc kubenswrapper[4815]: I0225 13:40:20.101215 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerName="ceilometer-notification-agent" containerID="cri-o://5df069ce02edb3750aaa11eb2126eb7af892e52967faa9f70cdd8a79bf0a1dfe" gracePeriod=30 Feb 25 13:40:20 crc kubenswrapper[4815]: I0225 13:40:20.425001 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 13:40:20 crc kubenswrapper[4815]: I0225 13:40:20.581707 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1407af2f-82a3-4ccb-9667-03d441ff34dc","Type":"ContainerStarted","Data":"800b371dbdc8ffff269d1bd672dc9fcf6a5310d38cceef187c71c9954b09603e"} Feb 25 13:40:20 crc kubenswrapper[4815]: I0225 13:40:20.585552 4815 generic.go:334] "Generic (PLEG): container finished" podID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerID="e7af682dbfbb6a00b57a0296b16bc311e4ee020e018576c68f0a1cc9b32f4c9b" exitCode=0 Feb 25 13:40:20 crc kubenswrapper[4815]: I0225 13:40:20.585586 4815 generic.go:334] "Generic (PLEG): container finished" podID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerID="379123921ac18dbae4ad9a5710dbbdfda9c066f6bde614a4ab7bd51a5ae51958" exitCode=2 Feb 25 13:40:20 crc kubenswrapper[4815]: I0225 13:40:20.585596 4815 generic.go:334] "Generic (PLEG): container finished" podID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerID="9400070e3a927941ad9839f3539cd29755bf156d96c7eb6767afd33b839509ad" exitCode=0 Feb 25 13:40:20 crc kubenswrapper[4815]: I0225 13:40:20.585555 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"922cea47-8f1c-4f9d-bbd4-f7a82c364356","Type":"ContainerDied","Data":"e7af682dbfbb6a00b57a0296b16bc311e4ee020e018576c68f0a1cc9b32f4c9b"} Feb 25 13:40:20 crc kubenswrapper[4815]: I0225 13:40:20.585631 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"922cea47-8f1c-4f9d-bbd4-f7a82c364356","Type":"ContainerDied","Data":"379123921ac18dbae4ad9a5710dbbdfda9c066f6bde614a4ab7bd51a5ae51958"} Feb 25 13:40:20 crc kubenswrapper[4815]: I0225 13:40:20.585646 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"922cea47-8f1c-4f9d-bbd4-f7a82c364356","Type":"ContainerDied","Data":"9400070e3a927941ad9839f3539cd29755bf156d96c7eb6767afd33b839509ad"} Feb 25 13:40:20 crc kubenswrapper[4815]: I0225 13:40:20.955082 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312" path="/var/lib/kubelet/pods/9d3ba3f7-a7b4-41c7-b39f-dcd6e28ed312/volumes" Feb 25 13:40:21 crc kubenswrapper[4815]: I0225 13:40:21.342131 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:40:21 crc kubenswrapper[4815]: I0225 13:40:21.342744 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:40:21 crc kubenswrapper[4815]: I0225 13:40:21.342822 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:40:21 crc kubenswrapper[4815]: I0225 13:40:21.344183 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"90119f271263ed85c7a4bae34f95ffae1fb69e1997807ded5781ce1fee85b5cc"} pod="openshift-machine-config-operator/machine-config-daemon-frc75" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 13:40:21 crc kubenswrapper[4815]: I0225 13:40:21.344271 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" containerID="cri-o://90119f271263ed85c7a4bae34f95ffae1fb69e1997807ded5781ce1fee85b5cc" gracePeriod=600 Feb 25 13:40:21 crc kubenswrapper[4815]: I0225 13:40:21.597697 4815 generic.go:334] "Generic (PLEG): container finished" podID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerID="90119f271263ed85c7a4bae34f95ffae1fb69e1997807ded5781ce1fee85b5cc" exitCode=0 Feb 25 13:40:21 crc kubenswrapper[4815]: I0225 13:40:21.597726 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerDied","Data":"90119f271263ed85c7a4bae34f95ffae1fb69e1997807ded5781ce1fee85b5cc"} Feb 25 13:40:21 crc kubenswrapper[4815]: I0225 13:40:21.598629 4815 scope.go:117] "RemoveContainer" containerID="65cc59f637c0af5e4fc8efb7df93ca963c3aa3c1788d5699757ce0a0287dce17" Feb 25 13:40:21 crc kubenswrapper[4815]: I0225 13:40:21.599171 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1407af2f-82a3-4ccb-9667-03d441ff34dc","Type":"ContainerStarted","Data":"6a9213c77557517422a2ef742c6b028763c3178bc95fa98338575ea821dbb62a"} Feb 25 13:40:21 crc kubenswrapper[4815]: I0225 13:40:21.599615 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 25 13:40:21 crc kubenswrapper[4815]: I0225 13:40:21.630772 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.256616201 podStartE2EDuration="2.630728567s" podCreationTimestamp="2026-02-25 13:40:19 +0000 UTC" firstStartedPulling="2026-02-25 13:40:20.432157526 +0000 UTC m=+1178.233255590" lastFinishedPulling="2026-02-25 13:40:20.806269862 +0000 UTC m=+1178.607367956" observedRunningTime="2026-02-25 13:40:21.617895898 +0000 UTC m=+1179.418993952" watchObservedRunningTime="2026-02-25 13:40:21.630728567 +0000 UTC m=+1179.431826621" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.173617 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.340317 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-config-data\") pod \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.340390 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-sg-core-conf-yaml\") pod \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.340488 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/922cea47-8f1c-4f9d-bbd4-f7a82c364356-log-httpd\") pod \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.340538 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jn2vt\" (UniqueName: \"kubernetes.io/projected/922cea47-8f1c-4f9d-bbd4-f7a82c364356-kube-api-access-jn2vt\") pod \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.340564 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-combined-ca-bundle\") pod \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.340582 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-scripts\") pod \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.340633 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/922cea47-8f1c-4f9d-bbd4-f7a82c364356-run-httpd\") pod \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\" (UID: \"922cea47-8f1c-4f9d-bbd4-f7a82c364356\") " Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.341033 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/922cea47-8f1c-4f9d-bbd4-f7a82c364356-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "922cea47-8f1c-4f9d-bbd4-f7a82c364356" (UID: "922cea47-8f1c-4f9d-bbd4-f7a82c364356"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.341208 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/922cea47-8f1c-4f9d-bbd4-f7a82c364356-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "922cea47-8f1c-4f9d-bbd4-f7a82c364356" (UID: "922cea47-8f1c-4f9d-bbd4-f7a82c364356"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.346098 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-scripts" (OuterVolumeSpecName: "scripts") pod "922cea47-8f1c-4f9d-bbd4-f7a82c364356" (UID: "922cea47-8f1c-4f9d-bbd4-f7a82c364356"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.358280 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/922cea47-8f1c-4f9d-bbd4-f7a82c364356-kube-api-access-jn2vt" (OuterVolumeSpecName: "kube-api-access-jn2vt") pod "922cea47-8f1c-4f9d-bbd4-f7a82c364356" (UID: "922cea47-8f1c-4f9d-bbd4-f7a82c364356"). InnerVolumeSpecName "kube-api-access-jn2vt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.370643 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "922cea47-8f1c-4f9d-bbd4-f7a82c364356" (UID: "922cea47-8f1c-4f9d-bbd4-f7a82c364356"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.409922 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "922cea47-8f1c-4f9d-bbd4-f7a82c364356" (UID: "922cea47-8f1c-4f9d-bbd4-f7a82c364356"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.452600 4815 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/922cea47-8f1c-4f9d-bbd4-f7a82c364356-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.452637 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jn2vt\" (UniqueName: \"kubernetes.io/projected/922cea47-8f1c-4f9d-bbd4-f7a82c364356-kube-api-access-jn2vt\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.452655 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.452667 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.452678 4815 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/922cea47-8f1c-4f9d-bbd4-f7a82c364356-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.452689 4815 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.453696 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-config-data" (OuterVolumeSpecName: "config-data") pod "922cea47-8f1c-4f9d-bbd4-f7a82c364356" (UID: "922cea47-8f1c-4f9d-bbd4-f7a82c364356"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.554367 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/922cea47-8f1c-4f9d-bbd4-f7a82c364356-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.609674 4815 generic.go:334] "Generic (PLEG): container finished" podID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerID="5df069ce02edb3750aaa11eb2126eb7af892e52967faa9f70cdd8a79bf0a1dfe" exitCode=0 Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.609735 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"922cea47-8f1c-4f9d-bbd4-f7a82c364356","Type":"ContainerDied","Data":"5df069ce02edb3750aaa11eb2126eb7af892e52967faa9f70cdd8a79bf0a1dfe"} Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.609765 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"922cea47-8f1c-4f9d-bbd4-f7a82c364356","Type":"ContainerDied","Data":"0f2aefb641682236d9d5f46024d7cfd956ac7ccea175a9213049a8ed5e79e9d5"} Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.609789 4815 scope.go:117] "RemoveContainer" containerID="e7af682dbfbb6a00b57a0296b16bc311e4ee020e018576c68f0a1cc9b32f4c9b" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.609901 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.626059 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerStarted","Data":"f21f73b0f5d16d6a34d568e8c21d1a82e30512997d6c1c659ca1e100017fbcdd"} Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.655270 4815 scope.go:117] "RemoveContainer" containerID="379123921ac18dbae4ad9a5710dbbdfda9c066f6bde614a4ab7bd51a5ae51958" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.674585 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.693299 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.704187 4815 scope.go:117] "RemoveContainer" containerID="5df069ce02edb3750aaa11eb2126eb7af892e52967faa9f70cdd8a79bf0a1dfe" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.708949 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:40:22 crc kubenswrapper[4815]: E0225 13:40:22.709558 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerName="proxy-httpd" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.709605 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerName="proxy-httpd" Feb 25 13:40:22 crc kubenswrapper[4815]: E0225 13:40:22.709638 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerName="sg-core" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.709647 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerName="sg-core" Feb 25 13:40:22 crc kubenswrapper[4815]: E0225 13:40:22.709667 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerName="ceilometer-central-agent" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.709675 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerName="ceilometer-central-agent" Feb 25 13:40:22 crc kubenswrapper[4815]: E0225 13:40:22.709694 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerName="ceilometer-notification-agent" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.709702 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerName="ceilometer-notification-agent" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.709918 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerName="proxy-httpd" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.709933 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerName="ceilometer-notification-agent" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.709947 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerName="sg-core" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.709975 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" containerName="ceilometer-central-agent" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.713320 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.718671 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.718912 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.719363 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.722778 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.729252 4815 scope.go:117] "RemoveContainer" containerID="9400070e3a927941ad9839f3539cd29755bf156d96c7eb6767afd33b839509ad" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.757024 4815 scope.go:117] "RemoveContainer" containerID="e7af682dbfbb6a00b57a0296b16bc311e4ee020e018576c68f0a1cc9b32f4c9b" Feb 25 13:40:22 crc kubenswrapper[4815]: E0225 13:40:22.757745 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7af682dbfbb6a00b57a0296b16bc311e4ee020e018576c68f0a1cc9b32f4c9b\": container with ID starting with e7af682dbfbb6a00b57a0296b16bc311e4ee020e018576c68f0a1cc9b32f4c9b not found: ID does not exist" containerID="e7af682dbfbb6a00b57a0296b16bc311e4ee020e018576c68f0a1cc9b32f4c9b" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.757802 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7af682dbfbb6a00b57a0296b16bc311e4ee020e018576c68f0a1cc9b32f4c9b"} err="failed to get container status \"e7af682dbfbb6a00b57a0296b16bc311e4ee020e018576c68f0a1cc9b32f4c9b\": rpc error: code = NotFound desc = could not find container \"e7af682dbfbb6a00b57a0296b16bc311e4ee020e018576c68f0a1cc9b32f4c9b\": container with ID starting with e7af682dbfbb6a00b57a0296b16bc311e4ee020e018576c68f0a1cc9b32f4c9b not found: ID does not exist" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.757837 4815 scope.go:117] "RemoveContainer" containerID="379123921ac18dbae4ad9a5710dbbdfda9c066f6bde614a4ab7bd51a5ae51958" Feb 25 13:40:22 crc kubenswrapper[4815]: E0225 13:40:22.758181 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"379123921ac18dbae4ad9a5710dbbdfda9c066f6bde614a4ab7bd51a5ae51958\": container with ID starting with 379123921ac18dbae4ad9a5710dbbdfda9c066f6bde614a4ab7bd51a5ae51958 not found: ID does not exist" containerID="379123921ac18dbae4ad9a5710dbbdfda9c066f6bde614a4ab7bd51a5ae51958" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.758221 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"379123921ac18dbae4ad9a5710dbbdfda9c066f6bde614a4ab7bd51a5ae51958"} err="failed to get container status \"379123921ac18dbae4ad9a5710dbbdfda9c066f6bde614a4ab7bd51a5ae51958\": rpc error: code = NotFound desc = could not find container \"379123921ac18dbae4ad9a5710dbbdfda9c066f6bde614a4ab7bd51a5ae51958\": container with ID starting with 379123921ac18dbae4ad9a5710dbbdfda9c066f6bde614a4ab7bd51a5ae51958 not found: ID does not exist" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.758247 4815 scope.go:117] "RemoveContainer" containerID="5df069ce02edb3750aaa11eb2126eb7af892e52967faa9f70cdd8a79bf0a1dfe" Feb 25 13:40:22 crc kubenswrapper[4815]: E0225 13:40:22.758618 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5df069ce02edb3750aaa11eb2126eb7af892e52967faa9f70cdd8a79bf0a1dfe\": container with ID starting with 5df069ce02edb3750aaa11eb2126eb7af892e52967faa9f70cdd8a79bf0a1dfe not found: ID does not exist" containerID="5df069ce02edb3750aaa11eb2126eb7af892e52967faa9f70cdd8a79bf0a1dfe" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.758675 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5df069ce02edb3750aaa11eb2126eb7af892e52967faa9f70cdd8a79bf0a1dfe"} err="failed to get container status \"5df069ce02edb3750aaa11eb2126eb7af892e52967faa9f70cdd8a79bf0a1dfe\": rpc error: code = NotFound desc = could not find container \"5df069ce02edb3750aaa11eb2126eb7af892e52967faa9f70cdd8a79bf0a1dfe\": container with ID starting with 5df069ce02edb3750aaa11eb2126eb7af892e52967faa9f70cdd8a79bf0a1dfe not found: ID does not exist" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.758723 4815 scope.go:117] "RemoveContainer" containerID="9400070e3a927941ad9839f3539cd29755bf156d96c7eb6767afd33b839509ad" Feb 25 13:40:22 crc kubenswrapper[4815]: E0225 13:40:22.759114 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9400070e3a927941ad9839f3539cd29755bf156d96c7eb6767afd33b839509ad\": container with ID starting with 9400070e3a927941ad9839f3539cd29755bf156d96c7eb6767afd33b839509ad not found: ID does not exist" containerID="9400070e3a927941ad9839f3539cd29755bf156d96c7eb6767afd33b839509ad" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.759155 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9400070e3a927941ad9839f3539cd29755bf156d96c7eb6767afd33b839509ad"} err="failed to get container status \"9400070e3a927941ad9839f3539cd29755bf156d96c7eb6767afd33b839509ad\": rpc error: code = NotFound desc = could not find container \"9400070e3a927941ad9839f3539cd29755bf156d96c7eb6767afd33b839509ad\": container with ID starting with 9400070e3a927941ad9839f3539cd29755bf156d96c7eb6767afd33b839509ad not found: ID does not exist" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.858939 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b019b447-6312-4672-861b-2692658b739d-log-httpd\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.858991 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b019b447-6312-4672-861b-2692658b739d-run-httpd\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.859122 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.859182 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.859342 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.859496 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-scripts\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.859567 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm9xh\" (UniqueName: \"kubernetes.io/projected/b019b447-6312-4672-861b-2692658b739d-kube-api-access-gm9xh\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.859802 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-config-data\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.946777 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="922cea47-8f1c-4f9d-bbd4-f7a82c364356" path="/var/lib/kubelet/pods/922cea47-8f1c-4f9d-bbd4-f7a82c364356/volumes" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.961841 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.961927 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-scripts\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.961960 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm9xh\" (UniqueName: \"kubernetes.io/projected/b019b447-6312-4672-861b-2692658b739d-kube-api-access-gm9xh\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.962045 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-config-data\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.962086 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b019b447-6312-4672-861b-2692658b739d-log-httpd\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.962131 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b019b447-6312-4672-861b-2692658b739d-run-httpd\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.962174 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.962751 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b019b447-6312-4672-861b-2692658b739d-log-httpd\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.962833 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b019b447-6312-4672-861b-2692658b739d-run-httpd\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.962931 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.967013 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-scripts\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.967398 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.968125 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.970041 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-config-data\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.971078 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.982139 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 25 13:40:22 crc kubenswrapper[4815]: I0225 13:40:22.982830 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm9xh\" (UniqueName: \"kubernetes.io/projected/b019b447-6312-4672-861b-2692658b739d-kube-api-access-gm9xh\") pod \"ceilometer-0\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " pod="openstack/ceilometer-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.044276 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.519882 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-xkz82"] Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.521857 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xkz82" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.526245 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.526323 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.529998 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.541355 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xkz82"] Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.651164 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b019b447-6312-4672-861b-2692658b739d","Type":"ContainerStarted","Data":"d8ca2a8d15e3cb90df3ebc5f1f20223e16f77f79499f8f34e7321381e58851ce"} Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.676114 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-scripts\") pod \"nova-cell0-cell-mapping-xkz82\" (UID: \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\") " pod="openstack/nova-cell0-cell-mapping-xkz82" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.676350 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-config-data\") pod \"nova-cell0-cell-mapping-xkz82\" (UID: \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\") " pod="openstack/nova-cell0-cell-mapping-xkz82" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.676443 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xkz82\" (UID: \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\") " pod="openstack/nova-cell0-cell-mapping-xkz82" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.676602 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzctv\" (UniqueName: \"kubernetes.io/projected/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-kube-api-access-wzctv\") pod \"nova-cell0-cell-mapping-xkz82\" (UID: \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\") " pod="openstack/nova-cell0-cell-mapping-xkz82" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.698596 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.699755 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.701374 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.710399 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.778276 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzctv\" (UniqueName: \"kubernetes.io/projected/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-kube-api-access-wzctv\") pod \"nova-cell0-cell-mapping-xkz82\" (UID: \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\") " pod="openstack/nova-cell0-cell-mapping-xkz82" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.778789 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-scripts\") pod \"nova-cell0-cell-mapping-xkz82\" (UID: \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\") " pod="openstack/nova-cell0-cell-mapping-xkz82" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.778938 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-config-data\") pod \"nova-cell0-cell-mapping-xkz82\" (UID: \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\") " pod="openstack/nova-cell0-cell-mapping-xkz82" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.778962 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xkz82\" (UID: \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\") " pod="openstack/nova-cell0-cell-mapping-xkz82" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.784072 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-scripts\") pod \"nova-cell0-cell-mapping-xkz82\" (UID: \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\") " pod="openstack/nova-cell0-cell-mapping-xkz82" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.786781 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xkz82\" (UID: \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\") " pod="openstack/nova-cell0-cell-mapping-xkz82" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.800950 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.805757 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.809684 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-config-data\") pod \"nova-cell0-cell-mapping-xkz82\" (UID: \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\") " pod="openstack/nova-cell0-cell-mapping-xkz82" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.810044 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.817130 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzctv\" (UniqueName: \"kubernetes.io/projected/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-kube-api-access-wzctv\") pod \"nova-cell0-cell-mapping-xkz82\" (UID: \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\") " pod="openstack/nova-cell0-cell-mapping-xkz82" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.823670 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.851611 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.859887 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.862995 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.892830 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.893742 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adaeacfe-023a-414e-a0c4-c59d8a181427-config-data\") pod \"nova-scheduler-0\" (UID: \"adaeacfe-023a-414e-a0c4-c59d8a181427\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.893850 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqdgw\" (UniqueName: \"kubernetes.io/projected/adaeacfe-023a-414e-a0c4-c59d8a181427-kube-api-access-dqdgw\") pod \"nova-scheduler-0\" (UID: \"adaeacfe-023a-414e-a0c4-c59d8a181427\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.894054 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adaeacfe-023a-414e-a0c4-c59d8a181427-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"adaeacfe-023a-414e-a0c4-c59d8a181427\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.905963 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xkz82" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.950941 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.952196 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.969010 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.996095 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\") " pod="openstack/nova-metadata-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.996408 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20edc318-3cd1-42a8-ac54-728c466e8e4c-logs\") pod \"nova-api-0\" (UID: \"20edc318-3cd1-42a8-ac54-728c466e8e4c\") " pod="openstack/nova-api-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.996437 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adaeacfe-023a-414e-a0c4-c59d8a181427-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"adaeacfe-023a-414e-a0c4-c59d8a181427\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.996457 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20edc318-3cd1-42a8-ac54-728c466e8e4c-config-data\") pod \"nova-api-0\" (UID: \"20edc318-3cd1-42a8-ac54-728c466e8e4c\") " pod="openstack/nova-api-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.996503 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-config-data\") pod \"nova-metadata-0\" (UID: \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\") " pod="openstack/nova-metadata-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.997716 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adaeacfe-023a-414e-a0c4-c59d8a181427-config-data\") pod \"nova-scheduler-0\" (UID: \"adaeacfe-023a-414e-a0c4-c59d8a181427\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.998251 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-logs\") pod \"nova-metadata-0\" (UID: \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\") " pod="openstack/nova-metadata-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.999669 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8nxx\" (UniqueName: \"kubernetes.io/projected/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-kube-api-access-v8nxx\") pod \"nova-metadata-0\" (UID: \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\") " pod="openstack/nova-metadata-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.999716 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j7sj\" (UniqueName: \"kubernetes.io/projected/20edc318-3cd1-42a8-ac54-728c466e8e4c-kube-api-access-5j7sj\") pod \"nova-api-0\" (UID: \"20edc318-3cd1-42a8-ac54-728c466e8e4c\") " pod="openstack/nova-api-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.999756 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqdgw\" (UniqueName: \"kubernetes.io/projected/adaeacfe-023a-414e-a0c4-c59d8a181427-kube-api-access-dqdgw\") pod \"nova-scheduler-0\" (UID: \"adaeacfe-023a-414e-a0c4-c59d8a181427\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:23 crc kubenswrapper[4815]: I0225 13:40:23.999805 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20edc318-3cd1-42a8-ac54-728c466e8e4c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"20edc318-3cd1-42a8-ac54-728c466e8e4c\") " pod="openstack/nova-api-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.004232 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adaeacfe-023a-414e-a0c4-c59d8a181427-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"adaeacfe-023a-414e-a0c4-c59d8a181427\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.004698 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adaeacfe-023a-414e-a0c4-c59d8a181427-config-data\") pod \"nova-scheduler-0\" (UID: \"adaeacfe-023a-414e-a0c4-c59d8a181427\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.039146 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqdgw\" (UniqueName: \"kubernetes.io/projected/adaeacfe-023a-414e-a0c4-c59d8a181427-kube-api-access-dqdgw\") pod \"nova-scheduler-0\" (UID: \"adaeacfe-023a-414e-a0c4-c59d8a181427\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.039211 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.053571 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-r49tp"] Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.055061 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.064620 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-r49tp"] Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.101061 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.101123 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-config-data\") pod \"nova-metadata-0\" (UID: \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\") " pod="openstack/nova-metadata-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.101160 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-logs\") pod \"nova-metadata-0\" (UID: \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\") " pod="openstack/nova-metadata-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.101181 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8nxx\" (UniqueName: \"kubernetes.io/projected/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-kube-api-access-v8nxx\") pod \"nova-metadata-0\" (UID: \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\") " pod="openstack/nova-metadata-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.101199 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j7sj\" (UniqueName: \"kubernetes.io/projected/20edc318-3cd1-42a8-ac54-728c466e8e4c-kube-api-access-5j7sj\") pod \"nova-api-0\" (UID: \"20edc318-3cd1-42a8-ac54-728c466e8e4c\") " pod="openstack/nova-api-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.101233 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20edc318-3cd1-42a8-ac54-728c466e8e4c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"20edc318-3cd1-42a8-ac54-728c466e8e4c\") " pod="openstack/nova-api-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.101277 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g966\" (UniqueName: \"kubernetes.io/projected/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-kube-api-access-7g966\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.101303 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\") " pod="openstack/nova-metadata-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.101353 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20edc318-3cd1-42a8-ac54-728c466e8e4c-logs\") pod \"nova-api-0\" (UID: \"20edc318-3cd1-42a8-ac54-728c466e8e4c\") " pod="openstack/nova-api-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.101371 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.101388 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20edc318-3cd1-42a8-ac54-728c466e8e4c-config-data\") pod \"nova-api-0\" (UID: \"20edc318-3cd1-42a8-ac54-728c466e8e4c\") " pod="openstack/nova-api-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.106174 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20edc318-3cd1-42a8-ac54-728c466e8e4c-logs\") pod \"nova-api-0\" (UID: \"20edc318-3cd1-42a8-ac54-728c466e8e4c\") " pod="openstack/nova-api-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.107112 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20edc318-3cd1-42a8-ac54-728c466e8e4c-config-data\") pod \"nova-api-0\" (UID: \"20edc318-3cd1-42a8-ac54-728c466e8e4c\") " pod="openstack/nova-api-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.107197 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-logs\") pod \"nova-metadata-0\" (UID: \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\") " pod="openstack/nova-metadata-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.111780 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-config-data\") pod \"nova-metadata-0\" (UID: \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\") " pod="openstack/nova-metadata-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.122078 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\") " pod="openstack/nova-metadata-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.122115 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20edc318-3cd1-42a8-ac54-728c466e8e4c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"20edc318-3cd1-42a8-ac54-728c466e8e4c\") " pod="openstack/nova-api-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.130963 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j7sj\" (UniqueName: \"kubernetes.io/projected/20edc318-3cd1-42a8-ac54-728c466e8e4c-kube-api-access-5j7sj\") pod \"nova-api-0\" (UID: \"20edc318-3cd1-42a8-ac54-728c466e8e4c\") " pod="openstack/nova-api-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.145031 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8nxx\" (UniqueName: \"kubernetes.io/projected/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-kube-api-access-v8nxx\") pod \"nova-metadata-0\" (UID: \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\") " pod="openstack/nova-metadata-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.203794 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.203855 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.203889 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g966\" (UniqueName: \"kubernetes.io/projected/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-kube-api-access-7g966\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.203921 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.203953 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-config\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.203975 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.204003 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.204018 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-dns-svc\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.204065 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9lpp\" (UniqueName: \"kubernetes.io/projected/437f9211-e28f-4da9-993b-9a4dd71796f6-kube-api-access-n9lpp\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.205095 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.208471 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.218055 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.227064 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g966\" (UniqueName: \"kubernetes.io/projected/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-kube-api-access-7g966\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.228351 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.305563 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.305643 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.305678 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-config\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.305712 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-dns-svc\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.305763 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9lpp\" (UniqueName: \"kubernetes.io/projected/437f9211-e28f-4da9-993b-9a4dd71796f6-kube-api-access-n9lpp\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.305815 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.306887 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.308021 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.308705 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.309300 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-config\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.309803 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-dns-svc\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.310496 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.331993 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9lpp\" (UniqueName: \"kubernetes.io/projected/437f9211-e28f-4da9-993b-9a4dd71796f6-kube-api-access-n9lpp\") pod \"dnsmasq-dns-865f5d856f-r49tp\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.332376 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.479445 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xkz82"] Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.620015 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.717807 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b019b447-6312-4672-861b-2692658b739d","Type":"ContainerStarted","Data":"d5e81f0eb1705fa7f0ce74497c1f61a949c0f6d42cbad1965f8c1e08a1ad7440"} Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.757369 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5n7jw"] Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.758500 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5n7jw" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.768872 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.769057 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.808586 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5n7jw"] Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.928465 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5n7jw\" (UID: \"98fc77c3-6063-41a5-ba29-0875703dfe05\") " pod="openstack/nova-cell1-conductor-db-sync-5n7jw" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.928551 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvhnv\" (UniqueName: \"kubernetes.io/projected/98fc77c3-6063-41a5-ba29-0875703dfe05-kube-api-access-nvhnv\") pod \"nova-cell1-conductor-db-sync-5n7jw\" (UID: \"98fc77c3-6063-41a5-ba29-0875703dfe05\") " pod="openstack/nova-cell1-conductor-db-sync-5n7jw" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.928580 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-scripts\") pod \"nova-cell1-conductor-db-sync-5n7jw\" (UID: \"98fc77c3-6063-41a5-ba29-0875703dfe05\") " pod="openstack/nova-cell1-conductor-db-sync-5n7jw" Feb 25 13:40:24 crc kubenswrapper[4815]: I0225 13:40:24.928692 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-config-data\") pod \"nova-cell1-conductor-db-sync-5n7jw\" (UID: \"98fc77c3-6063-41a5-ba29-0875703dfe05\") " pod="openstack/nova-cell1-conductor-db-sync-5n7jw" Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.031035 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-config-data\") pod \"nova-cell1-conductor-db-sync-5n7jw\" (UID: \"98fc77c3-6063-41a5-ba29-0875703dfe05\") " pod="openstack/nova-cell1-conductor-db-sync-5n7jw" Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.031105 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5n7jw\" (UID: \"98fc77c3-6063-41a5-ba29-0875703dfe05\") " pod="openstack/nova-cell1-conductor-db-sync-5n7jw" Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.031175 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvhnv\" (UniqueName: \"kubernetes.io/projected/98fc77c3-6063-41a5-ba29-0875703dfe05-kube-api-access-nvhnv\") pod \"nova-cell1-conductor-db-sync-5n7jw\" (UID: \"98fc77c3-6063-41a5-ba29-0875703dfe05\") " pod="openstack/nova-cell1-conductor-db-sync-5n7jw" Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.031228 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-scripts\") pod \"nova-cell1-conductor-db-sync-5n7jw\" (UID: \"98fc77c3-6063-41a5-ba29-0875703dfe05\") " pod="openstack/nova-cell1-conductor-db-sync-5n7jw" Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.039437 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-config-data\") pod \"nova-cell1-conductor-db-sync-5n7jw\" (UID: \"98fc77c3-6063-41a5-ba29-0875703dfe05\") " pod="openstack/nova-cell1-conductor-db-sync-5n7jw" Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.039561 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5n7jw\" (UID: \"98fc77c3-6063-41a5-ba29-0875703dfe05\") " pod="openstack/nova-cell1-conductor-db-sync-5n7jw" Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.039933 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-scripts\") pod \"nova-cell1-conductor-db-sync-5n7jw\" (UID: \"98fc77c3-6063-41a5-ba29-0875703dfe05\") " pod="openstack/nova-cell1-conductor-db-sync-5n7jw" Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.049313 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvhnv\" (UniqueName: \"kubernetes.io/projected/98fc77c3-6063-41a5-ba29-0875703dfe05-kube-api-access-nvhnv\") pod \"nova-cell1-conductor-db-sync-5n7jw\" (UID: \"98fc77c3-6063-41a5-ba29-0875703dfe05\") " pod="openstack/nova-cell1-conductor-db-sync-5n7jw" Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.104253 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5n7jw" Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.218136 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.235914 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 13:40:25 crc kubenswrapper[4815]: W0225 13:40:25.242656 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20edc318_3cd1_42a8_ac54_728c466e8e4c.slice/crio-f645b7292920975d94016d82de4c8de2586dffcf620c1d6a76fe5610f9b89494 WatchSource:0}: Error finding container f645b7292920975d94016d82de4c8de2586dffcf620c1d6a76fe5610f9b89494: Status 404 returned error can't find the container with id f645b7292920975d94016d82de4c8de2586dffcf620c1d6a76fe5610f9b89494 Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.248672 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 13:40:25 crc kubenswrapper[4815]: W0225 13:40:25.249952 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadaeacfe_023a_414e_a0c4_c59d8a181427.slice/crio-912c66e884cbc17db37b3073fc2bbeb101b6aff0d6e85e81e687c095e37d2700 WatchSource:0}: Error finding container 912c66e884cbc17db37b3073fc2bbeb101b6aff0d6e85e81e687c095e37d2700: Status 404 returned error can't find the container with id 912c66e884cbc17db37b3073fc2bbeb101b6aff0d6e85e81e687c095e37d2700 Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.256924 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.330073 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-r49tp"] Feb 25 13:40:25 crc kubenswrapper[4815]: W0225 13:40:25.351051 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod437f9211_e28f_4da9_993b_9a4dd71796f6.slice/crio-851e7f720f745965982651882048f37651e84c1620e69de60c30f8b474eabfcb WatchSource:0}: Error finding container 851e7f720f745965982651882048f37651e84c1620e69de60c30f8b474eabfcb: Status 404 returned error can't find the container with id 851e7f720f745965982651882048f37651e84c1620e69de60c30f8b474eabfcb Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.660185 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5n7jw"] Feb 25 13:40:25 crc kubenswrapper[4815]: W0225 13:40:25.694085 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98fc77c3_6063_41a5_ba29_0875703dfe05.slice/crio-5de7f8c3a0ef7bf87fa32c4b2e1f8b774034822051cff8e9fe29518bd31df415 WatchSource:0}: Error finding container 5de7f8c3a0ef7bf87fa32c4b2e1f8b774034822051cff8e9fe29518bd31df415: Status 404 returned error can't find the container with id 5de7f8c3a0ef7bf87fa32c4b2e1f8b774034822051cff8e9fe29518bd31df415 Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.731088 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518","Type":"ContainerStarted","Data":"09759eded458667cda35054112141f367d35475a6786b35cd280584efdfa155d"} Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.733483 4815 generic.go:334] "Generic (PLEG): container finished" podID="437f9211-e28f-4da9-993b-9a4dd71796f6" containerID="ac1b0e7571f7264de9e214a5d7804d337d6972277a4ca74433a716fbc1da94d9" exitCode=0 Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.733558 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-r49tp" event={"ID":"437f9211-e28f-4da9-993b-9a4dd71796f6","Type":"ContainerDied","Data":"ac1b0e7571f7264de9e214a5d7804d337d6972277a4ca74433a716fbc1da94d9"} Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.733585 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-r49tp" event={"ID":"437f9211-e28f-4da9-993b-9a4dd71796f6","Type":"ContainerStarted","Data":"851e7f720f745965982651882048f37651e84c1620e69de60c30f8b474eabfcb"} Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.735225 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xkz82" event={"ID":"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be","Type":"ContainerStarted","Data":"53c66837ec255d39fab425ff76173ac7ffa456950e2714481fd0aa74c95dbe5e"} Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.735267 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xkz82" event={"ID":"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be","Type":"ContainerStarted","Data":"02dbaf30114e5f57f4c51cfe74e2a7ece995922442d1c3e9ac59801c0e27b982"} Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.743543 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b019b447-6312-4672-861b-2692658b739d","Type":"ContainerStarted","Data":"5c6a67efff1d2c98e92ccb3c3727b0819482bb0d62c7f86fe25d2b1431ec54fa"} Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.746360 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5n7jw" event={"ID":"98fc77c3-6063-41a5-ba29-0875703dfe05","Type":"ContainerStarted","Data":"5de7f8c3a0ef7bf87fa32c4b2e1f8b774034822051cff8e9fe29518bd31df415"} Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.748295 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"52a88435-c7aa-4297-b4aa-ffcb6d1b8629","Type":"ContainerStarted","Data":"a68ecdeafad647a2be9dd60c56397004df4035b4d77c18b51a8bafbd239888c9"} Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.766352 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"adaeacfe-023a-414e-a0c4-c59d8a181427","Type":"ContainerStarted","Data":"912c66e884cbc17db37b3073fc2bbeb101b6aff0d6e85e81e687c095e37d2700"} Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.773637 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20edc318-3cd1-42a8-ac54-728c466e8e4c","Type":"ContainerStarted","Data":"f645b7292920975d94016d82de4c8de2586dffcf620c1d6a76fe5610f9b89494"} Feb 25 13:40:25 crc kubenswrapper[4815]: I0225 13:40:25.791244 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-xkz82" podStartSLOduration=2.791225943 podStartE2EDuration="2.791225943s" podCreationTimestamp="2026-02-25 13:40:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:40:25.7690252 +0000 UTC m=+1183.570123264" watchObservedRunningTime="2026-02-25 13:40:25.791225943 +0000 UTC m=+1183.592323997" Feb 25 13:40:26 crc kubenswrapper[4815]: I0225 13:40:26.788470 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5n7jw" event={"ID":"98fc77c3-6063-41a5-ba29-0875703dfe05","Type":"ContainerStarted","Data":"3b3fadc2c41679dd6cbd8de18d1e4ee943756cb1e45c6b8ea7320aee646486d0"} Feb 25 13:40:26 crc kubenswrapper[4815]: I0225 13:40:26.796291 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-r49tp" event={"ID":"437f9211-e28f-4da9-993b-9a4dd71796f6","Type":"ContainerStarted","Data":"9c8e4523cc93cae849c9317f02d216c204cd816fb738b3b6ead4dd808da01f5e"} Feb 25 13:40:26 crc kubenswrapper[4815]: I0225 13:40:26.796431 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:26 crc kubenswrapper[4815]: I0225 13:40:26.798768 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b019b447-6312-4672-861b-2692658b739d","Type":"ContainerStarted","Data":"9dcbabae071a33cf72605cd940a5e2af845afd5a2f96afa808607ec367d5037a"} Feb 25 13:40:26 crc kubenswrapper[4815]: I0225 13:40:26.818270 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-5n7jw" podStartSLOduration=2.818249795 podStartE2EDuration="2.818249795s" podCreationTimestamp="2026-02-25 13:40:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:40:26.805121907 +0000 UTC m=+1184.606219961" watchObservedRunningTime="2026-02-25 13:40:26.818249795 +0000 UTC m=+1184.619347839" Feb 25 13:40:26 crc kubenswrapper[4815]: I0225 13:40:26.830461 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-865f5d856f-r49tp" podStartSLOduration=3.830444785 podStartE2EDuration="3.830444785s" podCreationTimestamp="2026-02-25 13:40:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:40:26.820643347 +0000 UTC m=+1184.621741401" watchObservedRunningTime="2026-02-25 13:40:26.830444785 +0000 UTC m=+1184.631542839" Feb 25 13:40:27 crc kubenswrapper[4815]: I0225 13:40:27.534000 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 13:40:27 crc kubenswrapper[4815]: I0225 13:40:27.542028 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:40:29 crc kubenswrapper[4815]: I0225 13:40:29.844327 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"52a88435-c7aa-4297-b4aa-ffcb6d1b8629","Type":"ContainerStarted","Data":"e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca"} Feb 25 13:40:29 crc kubenswrapper[4815]: I0225 13:40:29.844418 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="52a88435-c7aa-4297-b4aa-ffcb6d1b8629" containerName="nova-metadata-metadata" containerID="cri-o://e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca" gracePeriod=30 Feb 25 13:40:29 crc kubenswrapper[4815]: I0225 13:40:29.844690 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="52a88435-c7aa-4297-b4aa-ffcb6d1b8629" containerName="nova-metadata-log" containerID="cri-o://4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d" gracePeriod=30 Feb 25 13:40:29 crc kubenswrapper[4815]: I0225 13:40:29.844884 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"52a88435-c7aa-4297-b4aa-ffcb6d1b8629","Type":"ContainerStarted","Data":"4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d"} Feb 25 13:40:29 crc kubenswrapper[4815]: I0225 13:40:29.848133 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"adaeacfe-023a-414e-a0c4-c59d8a181427","Type":"ContainerStarted","Data":"7178c331681a8ef2a0963b41b61126e06059b346fa5d5c111f6889cfb4923472"} Feb 25 13:40:29 crc kubenswrapper[4815]: I0225 13:40:29.852777 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20edc318-3cd1-42a8-ac54-728c466e8e4c","Type":"ContainerStarted","Data":"c155b9b7d6d0ab409fbfb22530a8d29bf9183f4ca91ebc5cc0433c864eefabdd"} Feb 25 13:40:29 crc kubenswrapper[4815]: I0225 13:40:29.852817 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20edc318-3cd1-42a8-ac54-728c466e8e4c","Type":"ContainerStarted","Data":"2012d1a4dc530731bf43543bc21dd983c7a3405bbd77e525c7e19653d98467ea"} Feb 25 13:40:29 crc kubenswrapper[4815]: I0225 13:40:29.855003 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518","Type":"ContainerStarted","Data":"8585f19dc562745bbf47de77c1c3d99029b63418cb239cffbf7aedaa87fb5fe1"} Feb 25 13:40:29 crc kubenswrapper[4815]: I0225 13:40:29.855050 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="fd374bbd-13c7-4b9c-91f9-6dfcce5cf518" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://8585f19dc562745bbf47de77c1c3d99029b63418cb239cffbf7aedaa87fb5fe1" gracePeriod=30 Feb 25 13:40:29 crc kubenswrapper[4815]: I0225 13:40:29.861431 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b019b447-6312-4672-861b-2692658b739d","Type":"ContainerStarted","Data":"8950b7b5ab3b75b777743934b8423157e85e74526e0934fe3a56d59911cc25d8"} Feb 25 13:40:29 crc kubenswrapper[4815]: I0225 13:40:29.861613 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 13:40:29 crc kubenswrapper[4815]: I0225 13:40:29.869851 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.460336899 podStartE2EDuration="6.869826867s" podCreationTimestamp="2026-02-25 13:40:23 +0000 UTC" firstStartedPulling="2026-02-25 13:40:25.269470492 +0000 UTC m=+1183.070568536" lastFinishedPulling="2026-02-25 13:40:28.67896045 +0000 UTC m=+1186.480058504" observedRunningTime="2026-02-25 13:40:29.866487986 +0000 UTC m=+1187.667586050" watchObservedRunningTime="2026-02-25 13:40:29.869826867 +0000 UTC m=+1187.670924921" Feb 25 13:40:29 crc kubenswrapper[4815]: I0225 13:40:29.907354 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.4728738789999998 podStartE2EDuration="6.907335814s" podCreationTimestamp="2026-02-25 13:40:23 +0000 UTC" firstStartedPulling="2026-02-25 13:40:25.247683732 +0000 UTC m=+1183.048781776" lastFinishedPulling="2026-02-25 13:40:28.682145657 +0000 UTC m=+1186.483243711" observedRunningTime="2026-02-25 13:40:29.906931182 +0000 UTC m=+1187.708029246" watchObservedRunningTime="2026-02-25 13:40:29.907335814 +0000 UTC m=+1187.708433868" Feb 25 13:40:29 crc kubenswrapper[4815]: I0225 13:40:29.907808 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.495762693 podStartE2EDuration="6.907798898s" podCreationTimestamp="2026-02-25 13:40:23 +0000 UTC" firstStartedPulling="2026-02-25 13:40:25.241940908 +0000 UTC m=+1183.043038962" lastFinishedPulling="2026-02-25 13:40:28.653977103 +0000 UTC m=+1186.455075167" observedRunningTime="2026-02-25 13:40:29.890055451 +0000 UTC m=+1187.691153505" watchObservedRunningTime="2026-02-25 13:40:29.907798898 +0000 UTC m=+1187.708896962" Feb 25 13:40:29 crc kubenswrapper[4815]: I0225 13:40:29.943236 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.796723636 podStartE2EDuration="7.943217272s" podCreationTimestamp="2026-02-25 13:40:22 +0000 UTC" firstStartedPulling="2026-02-25 13:40:23.533403573 +0000 UTC m=+1181.334501627" lastFinishedPulling="2026-02-25 13:40:28.679897209 +0000 UTC m=+1186.480995263" observedRunningTime="2026-02-25 13:40:29.927885527 +0000 UTC m=+1187.728983581" watchObservedRunningTime="2026-02-25 13:40:29.943217272 +0000 UTC m=+1187.744315326" Feb 25 13:40:29 crc kubenswrapper[4815]: I0225 13:40:29.951483 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.525196475 podStartE2EDuration="6.951462911s" podCreationTimestamp="2026-02-25 13:40:23 +0000 UTC" firstStartedPulling="2026-02-25 13:40:25.253680504 +0000 UTC m=+1183.054778558" lastFinishedPulling="2026-02-25 13:40:28.67994695 +0000 UTC m=+1186.481044994" observedRunningTime="2026-02-25 13:40:29.949057529 +0000 UTC m=+1187.750155583" watchObservedRunningTime="2026-02-25 13:40:29.951462911 +0000 UTC m=+1187.752560965" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.029040 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.466953 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.552934 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8nxx\" (UniqueName: \"kubernetes.io/projected/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-kube-api-access-v8nxx\") pod \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\" (UID: \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\") " Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.553029 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-combined-ca-bundle\") pod \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\" (UID: \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\") " Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.553071 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-config-data\") pod \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\" (UID: \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\") " Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.553151 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-logs\") pod \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\" (UID: \"52a88435-c7aa-4297-b4aa-ffcb6d1b8629\") " Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.554097 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-logs" (OuterVolumeSpecName: "logs") pod "52a88435-c7aa-4297-b4aa-ffcb6d1b8629" (UID: "52a88435-c7aa-4297-b4aa-ffcb6d1b8629"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.563656 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-kube-api-access-v8nxx" (OuterVolumeSpecName: "kube-api-access-v8nxx") pod "52a88435-c7aa-4297-b4aa-ffcb6d1b8629" (UID: "52a88435-c7aa-4297-b4aa-ffcb6d1b8629"). InnerVolumeSpecName "kube-api-access-v8nxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.598235 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-config-data" (OuterVolumeSpecName: "config-data") pod "52a88435-c7aa-4297-b4aa-ffcb6d1b8629" (UID: "52a88435-c7aa-4297-b4aa-ffcb6d1b8629"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.607606 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "52a88435-c7aa-4297-b4aa-ffcb6d1b8629" (UID: "52a88435-c7aa-4297-b4aa-ffcb6d1b8629"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.655848 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8nxx\" (UniqueName: \"kubernetes.io/projected/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-kube-api-access-v8nxx\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.655897 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.655918 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.655936 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52a88435-c7aa-4297-b4aa-ffcb6d1b8629-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.873803 4815 generic.go:334] "Generic (PLEG): container finished" podID="52a88435-c7aa-4297-b4aa-ffcb6d1b8629" containerID="e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca" exitCode=0 Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.874143 4815 generic.go:334] "Generic (PLEG): container finished" podID="52a88435-c7aa-4297-b4aa-ffcb6d1b8629" containerID="4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d" exitCode=143 Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.873888 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"52a88435-c7aa-4297-b4aa-ffcb6d1b8629","Type":"ContainerDied","Data":"e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca"} Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.873863 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.874250 4815 scope.go:117] "RemoveContainer" containerID="e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.874237 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"52a88435-c7aa-4297-b4aa-ffcb6d1b8629","Type":"ContainerDied","Data":"4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d"} Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.874323 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"52a88435-c7aa-4297-b4aa-ffcb6d1b8629","Type":"ContainerDied","Data":"a68ecdeafad647a2be9dd60c56397004df4035b4d77c18b51a8bafbd239888c9"} Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.903319 4815 scope.go:117] "RemoveContainer" containerID="4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.927450 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.943203 4815 scope.go:117] "RemoveContainer" containerID="e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca" Feb 25 13:40:30 crc kubenswrapper[4815]: E0225 13:40:30.951031 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca\": container with ID starting with e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca not found: ID does not exist" containerID="e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.951158 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca"} err="failed to get container status \"e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca\": rpc error: code = NotFound desc = could not find container \"e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca\": container with ID starting with e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca not found: ID does not exist" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.951193 4815 scope.go:117] "RemoveContainer" containerID="4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d" Feb 25 13:40:30 crc kubenswrapper[4815]: E0225 13:40:30.951769 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d\": container with ID starting with 4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d not found: ID does not exist" containerID="4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.951817 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d"} err="failed to get container status \"4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d\": rpc error: code = NotFound desc = could not find container \"4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d\": container with ID starting with 4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d not found: ID does not exist" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.951840 4815 scope.go:117] "RemoveContainer" containerID="e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.952213 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca"} err="failed to get container status \"e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca\": rpc error: code = NotFound desc = could not find container \"e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca\": container with ID starting with e6d42ff78eee5dbc50cb49ebba20de8421d255c78ac8cba83745bb662a56b8ca not found: ID does not exist" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.952262 4815 scope.go:117] "RemoveContainer" containerID="4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.952856 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d"} err="failed to get container status \"4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d\": rpc error: code = NotFound desc = could not find container \"4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d\": container with ID starting with 4a756428918592c7baa517853c30d74fe737336fd971a0653c157aaa4b807c8d not found: ID does not exist" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.980303 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.980359 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:40:30 crc kubenswrapper[4815]: E0225 13:40:30.981195 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52a88435-c7aa-4297-b4aa-ffcb6d1b8629" containerName="nova-metadata-log" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.981222 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="52a88435-c7aa-4297-b4aa-ffcb6d1b8629" containerName="nova-metadata-log" Feb 25 13:40:30 crc kubenswrapper[4815]: E0225 13:40:30.981286 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52a88435-c7aa-4297-b4aa-ffcb6d1b8629" containerName="nova-metadata-metadata" Feb 25 13:40:30 crc kubenswrapper[4815]: I0225 13:40:30.981297 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="52a88435-c7aa-4297-b4aa-ffcb6d1b8629" containerName="nova-metadata-metadata" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.006498 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="52a88435-c7aa-4297-b4aa-ffcb6d1b8629" containerName="nova-metadata-log" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.006660 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="52a88435-c7aa-4297-b4aa-ffcb6d1b8629" containerName="nova-metadata-metadata" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.012136 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.012234 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.014139 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.014327 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.177759 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " pod="openstack/nova-metadata-0" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.177865 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " pod="openstack/nova-metadata-0" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.177923 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-config-data\") pod \"nova-metadata-0\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " pod="openstack/nova-metadata-0" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.178019 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-logs\") pod \"nova-metadata-0\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " pod="openstack/nova-metadata-0" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.178040 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px2qf\" (UniqueName: \"kubernetes.io/projected/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-kube-api-access-px2qf\") pod \"nova-metadata-0\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " pod="openstack/nova-metadata-0" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.279183 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " pod="openstack/nova-metadata-0" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.279261 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " pod="openstack/nova-metadata-0" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.279299 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-config-data\") pod \"nova-metadata-0\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " pod="openstack/nova-metadata-0" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.279365 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px2qf\" (UniqueName: \"kubernetes.io/projected/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-kube-api-access-px2qf\") pod \"nova-metadata-0\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " pod="openstack/nova-metadata-0" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.279381 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-logs\") pod \"nova-metadata-0\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " pod="openstack/nova-metadata-0" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.279821 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-logs\") pod \"nova-metadata-0\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " pod="openstack/nova-metadata-0" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.289682 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " pod="openstack/nova-metadata-0" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.290128 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-config-data\") pod \"nova-metadata-0\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " pod="openstack/nova-metadata-0" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.290630 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " pod="openstack/nova-metadata-0" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.306189 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px2qf\" (UniqueName: \"kubernetes.io/projected/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-kube-api-access-px2qf\") pod \"nova-metadata-0\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " pod="openstack/nova-metadata-0" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.330106 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.804278 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:40:31 crc kubenswrapper[4815]: I0225 13:40:31.882982 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6","Type":"ContainerStarted","Data":"498640fed436778f248eabe37882517d60bfae2afb5e5b6f8035d2abd386398c"} Feb 25 13:40:32 crc kubenswrapper[4815]: I0225 13:40:32.896561 4815 generic.go:334] "Generic (PLEG): container finished" podID="947c26ca-d609-4a2a-a4d8-0ab4ae3f69be" containerID="53c66837ec255d39fab425ff76173ac7ffa456950e2714481fd0aa74c95dbe5e" exitCode=0 Feb 25 13:40:32 crc kubenswrapper[4815]: I0225 13:40:32.896662 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xkz82" event={"ID":"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be","Type":"ContainerDied","Data":"53c66837ec255d39fab425ff76173ac7ffa456950e2714481fd0aa74c95dbe5e"} Feb 25 13:40:32 crc kubenswrapper[4815]: I0225 13:40:32.898679 4815 generic.go:334] "Generic (PLEG): container finished" podID="98fc77c3-6063-41a5-ba29-0875703dfe05" containerID="3b3fadc2c41679dd6cbd8de18d1e4ee943756cb1e45c6b8ea7320aee646486d0" exitCode=0 Feb 25 13:40:32 crc kubenswrapper[4815]: I0225 13:40:32.898742 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5n7jw" event={"ID":"98fc77c3-6063-41a5-ba29-0875703dfe05","Type":"ContainerDied","Data":"3b3fadc2c41679dd6cbd8de18d1e4ee943756cb1e45c6b8ea7320aee646486d0"} Feb 25 13:40:32 crc kubenswrapper[4815]: I0225 13:40:32.901236 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6","Type":"ContainerStarted","Data":"8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73"} Feb 25 13:40:32 crc kubenswrapper[4815]: I0225 13:40:32.901273 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6","Type":"ContainerStarted","Data":"b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd"} Feb 25 13:40:32 crc kubenswrapper[4815]: I0225 13:40:32.951120 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52a88435-c7aa-4297-b4aa-ffcb6d1b8629" path="/var/lib/kubelet/pods/52a88435-c7aa-4297-b4aa-ffcb6d1b8629/volumes" Feb 25 13:40:32 crc kubenswrapper[4815]: I0225 13:40:32.959245 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.959212265 podStartE2EDuration="2.959212265s" podCreationTimestamp="2026-02-25 13:40:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:40:32.942751076 +0000 UTC m=+1190.743849170" watchObservedRunningTime="2026-02-25 13:40:32.959212265 +0000 UTC m=+1190.760310339" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.207100 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.208892 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.311469 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.332627 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.332687 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.373485 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.398797 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xkz82" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.406178 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5n7jw" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.559819 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-config-data\") pod \"98fc77c3-6063-41a5-ba29-0875703dfe05\" (UID: \"98fc77c3-6063-41a5-ba29-0875703dfe05\") " Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.559955 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-scripts\") pod \"98fc77c3-6063-41a5-ba29-0875703dfe05\" (UID: \"98fc77c3-6063-41a5-ba29-0875703dfe05\") " Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.560162 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-scripts\") pod \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\" (UID: \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\") " Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.560259 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-config-data\") pod \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\" (UID: \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\") " Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.560354 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvhnv\" (UniqueName: \"kubernetes.io/projected/98fc77c3-6063-41a5-ba29-0875703dfe05-kube-api-access-nvhnv\") pod \"98fc77c3-6063-41a5-ba29-0875703dfe05\" (UID: \"98fc77c3-6063-41a5-ba29-0875703dfe05\") " Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.560424 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-combined-ca-bundle\") pod \"98fc77c3-6063-41a5-ba29-0875703dfe05\" (UID: \"98fc77c3-6063-41a5-ba29-0875703dfe05\") " Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.560477 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzctv\" (UniqueName: \"kubernetes.io/projected/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-kube-api-access-wzctv\") pod \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\" (UID: \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\") " Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.560597 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-combined-ca-bundle\") pod \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\" (UID: \"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be\") " Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.566412 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98fc77c3-6063-41a5-ba29-0875703dfe05-kube-api-access-nvhnv" (OuterVolumeSpecName: "kube-api-access-nvhnv") pod "98fc77c3-6063-41a5-ba29-0875703dfe05" (UID: "98fc77c3-6063-41a5-ba29-0875703dfe05"). InnerVolumeSpecName "kube-api-access-nvhnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.577986 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-scripts" (OuterVolumeSpecName: "scripts") pod "947c26ca-d609-4a2a-a4d8-0ab4ae3f69be" (UID: "947c26ca-d609-4a2a-a4d8-0ab4ae3f69be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.577997 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-scripts" (OuterVolumeSpecName: "scripts") pod "98fc77c3-6063-41a5-ba29-0875703dfe05" (UID: "98fc77c3-6063-41a5-ba29-0875703dfe05"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.578237 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-kube-api-access-wzctv" (OuterVolumeSpecName: "kube-api-access-wzctv") pod "947c26ca-d609-4a2a-a4d8-0ab4ae3f69be" (UID: "947c26ca-d609-4a2a-a4d8-0ab4ae3f69be"). InnerVolumeSpecName "kube-api-access-wzctv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.588944 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-config-data" (OuterVolumeSpecName: "config-data") pod "947c26ca-d609-4a2a-a4d8-0ab4ae3f69be" (UID: "947c26ca-d609-4a2a-a4d8-0ab4ae3f69be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.595907 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "947c26ca-d609-4a2a-a4d8-0ab4ae3f69be" (UID: "947c26ca-d609-4a2a-a4d8-0ab4ae3f69be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.608712 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-config-data" (OuterVolumeSpecName: "config-data") pod "98fc77c3-6063-41a5-ba29-0875703dfe05" (UID: "98fc77c3-6063-41a5-ba29-0875703dfe05"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.614713 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98fc77c3-6063-41a5-ba29-0875703dfe05" (UID: "98fc77c3-6063-41a5-ba29-0875703dfe05"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.621752 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.667530 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.669064 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.674577 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.674779 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.674891 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.674979 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvhnv\" (UniqueName: \"kubernetes.io/projected/98fc77c3-6063-41a5-ba29-0875703dfe05-kube-api-access-nvhnv\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.675072 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98fc77c3-6063-41a5-ba29-0875703dfe05-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.675148 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzctv\" (UniqueName: \"kubernetes.io/projected/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be-kube-api-access-wzctv\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.719720 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2wmhn"] Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.719989 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" podUID="2e02f6a9-4025-45bc-a523-17b0c0758268" containerName="dnsmasq-dns" containerID="cri-o://f2077ba5aa6728327f15f64309d6fecfec0e0dd7553e042a47e119e5d1a5a998" gracePeriod=10 Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.936729 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xkz82" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.953424 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xkz82" event={"ID":"947c26ca-d609-4a2a-a4d8-0ab4ae3f69be","Type":"ContainerDied","Data":"02dbaf30114e5f57f4c51cfe74e2a7ece995922442d1c3e9ac59801c0e27b982"} Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.953463 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02dbaf30114e5f57f4c51cfe74e2a7ece995922442d1c3e9ac59801c0e27b982" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.961622 4815 generic.go:334] "Generic (PLEG): container finished" podID="2e02f6a9-4025-45bc-a523-17b0c0758268" containerID="f2077ba5aa6728327f15f64309d6fecfec0e0dd7553e042a47e119e5d1a5a998" exitCode=0 Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.961737 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" event={"ID":"2e02f6a9-4025-45bc-a523-17b0c0758268","Type":"ContainerDied","Data":"f2077ba5aa6728327f15f64309d6fecfec0e0dd7553e042a47e119e5d1a5a998"} Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.976329 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5n7jw" Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.980707 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5n7jw" event={"ID":"98fc77c3-6063-41a5-ba29-0875703dfe05","Type":"ContainerDied","Data":"5de7f8c3a0ef7bf87fa32c4b2e1f8b774034822051cff8e9fe29518bd31df415"} Feb 25 13:40:34 crc kubenswrapper[4815]: I0225 13:40:34.980761 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5de7f8c3a0ef7bf87fa32c4b2e1f8b774034822051cff8e9fe29518bd31df415" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.023001 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.054566 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.094146 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 13:40:35 crc kubenswrapper[4815]: E0225 13:40:35.094592 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="947c26ca-d609-4a2a-a4d8-0ab4ae3f69be" containerName="nova-manage" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.094609 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="947c26ca-d609-4a2a-a4d8-0ab4ae3f69be" containerName="nova-manage" Feb 25 13:40:35 crc kubenswrapper[4815]: E0225 13:40:35.094644 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98fc77c3-6063-41a5-ba29-0875703dfe05" containerName="nova-cell1-conductor-db-sync" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.094651 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="98fc77c3-6063-41a5-ba29-0875703dfe05" containerName="nova-cell1-conductor-db-sync" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.094815 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="98fc77c3-6063-41a5-ba29-0875703dfe05" containerName="nova-cell1-conductor-db-sync" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.094831 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="947c26ca-d609-4a2a-a4d8-0ab4ae3f69be" containerName="nova-manage" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.096419 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.101421 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.124369 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.138163 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.145005 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.145219 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6" containerName="nova-metadata-log" containerID="cri-o://b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd" gracePeriod=30 Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.145691 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6" containerName="nova-metadata-metadata" containerID="cri-o://8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73" gracePeriod=30 Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.185438 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501\") " pod="openstack/nova-cell1-conductor-0" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.185519 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45959\" (UniqueName: \"kubernetes.io/projected/3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501-kube-api-access-45959\") pod \"nova-cell1-conductor-0\" (UID: \"3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501\") " pod="openstack/nova-cell1-conductor-0" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.185544 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501\") " pod="openstack/nova-cell1-conductor-0" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.213558 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.250066 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="20edc318-3cd1-42a8-ac54-728c466e8e4c" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.199:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.286925 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-dns-swift-storage-0\") pod \"2e02f6a9-4025-45bc-a523-17b0c0758268\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.287026 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-ovsdbserver-sb\") pod \"2e02f6a9-4025-45bc-a523-17b0c0758268\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.287093 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-config\") pod \"2e02f6a9-4025-45bc-a523-17b0c0758268\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.287155 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-dns-svc\") pod \"2e02f6a9-4025-45bc-a523-17b0c0758268\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.287212 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-ovsdbserver-nb\") pod \"2e02f6a9-4025-45bc-a523-17b0c0758268\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.287245 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cr7gg\" (UniqueName: \"kubernetes.io/projected/2e02f6a9-4025-45bc-a523-17b0c0758268-kube-api-access-cr7gg\") pod \"2e02f6a9-4025-45bc-a523-17b0c0758268\" (UID: \"2e02f6a9-4025-45bc-a523-17b0c0758268\") " Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.287482 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501\") " pod="openstack/nova-cell1-conductor-0" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.287532 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45959\" (UniqueName: \"kubernetes.io/projected/3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501-kube-api-access-45959\") pod \"nova-cell1-conductor-0\" (UID: \"3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501\") " pod="openstack/nova-cell1-conductor-0" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.287551 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501\") " pod="openstack/nova-cell1-conductor-0" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.293992 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="20edc318-3cd1-42a8-ac54-728c466e8e4c" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.199:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.296548 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501\") " pod="openstack/nova-cell1-conductor-0" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.298646 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501\") " pod="openstack/nova-cell1-conductor-0" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.305693 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e02f6a9-4025-45bc-a523-17b0c0758268-kube-api-access-cr7gg" (OuterVolumeSpecName: "kube-api-access-cr7gg") pod "2e02f6a9-4025-45bc-a523-17b0c0758268" (UID: "2e02f6a9-4025-45bc-a523-17b0c0758268"). InnerVolumeSpecName "kube-api-access-cr7gg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.323982 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45959\" (UniqueName: \"kubernetes.io/projected/3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501-kube-api-access-45959\") pod \"nova-cell1-conductor-0\" (UID: \"3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501\") " pod="openstack/nova-cell1-conductor-0" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.353012 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-config" (OuterVolumeSpecName: "config") pod "2e02f6a9-4025-45bc-a523-17b0c0758268" (UID: "2e02f6a9-4025-45bc-a523-17b0c0758268"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.354044 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2e02f6a9-4025-45bc-a523-17b0c0758268" (UID: "2e02f6a9-4025-45bc-a523-17b0c0758268"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.359694 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2e02f6a9-4025-45bc-a523-17b0c0758268" (UID: "2e02f6a9-4025-45bc-a523-17b0c0758268"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.368854 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2e02f6a9-4025-45bc-a523-17b0c0758268" (UID: "2e02f6a9-4025-45bc-a523-17b0c0758268"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.380734 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2e02f6a9-4025-45bc-a523-17b0c0758268" (UID: "2e02f6a9-4025-45bc-a523-17b0c0758268"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.389764 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.389800 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.389811 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.389822 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cr7gg\" (UniqueName: \"kubernetes.io/projected/2e02f6a9-4025-45bc-a523-17b0c0758268-kube-api-access-cr7gg\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.389834 4815 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.389842 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2e02f6a9-4025-45bc-a523-17b0c0758268-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.430404 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.714573 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.802133 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-config-data\") pod \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.802187 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px2qf\" (UniqueName: \"kubernetes.io/projected/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-kube-api-access-px2qf\") pod \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.802214 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-logs\") pod \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.802252 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-nova-metadata-tls-certs\") pod \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.802450 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-combined-ca-bundle\") pod \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\" (UID: \"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6\") " Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.803637 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-logs" (OuterVolumeSpecName: "logs") pod "1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6" (UID: "1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.824926 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-kube-api-access-px2qf" (OuterVolumeSpecName: "kube-api-access-px2qf") pod "1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6" (UID: "1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6"). InnerVolumeSpecName "kube-api-access-px2qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.876681 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-config-data" (OuterVolumeSpecName: "config-data") pod "1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6" (UID: "1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.904570 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.904599 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px2qf\" (UniqueName: \"kubernetes.io/projected/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-kube-api-access-px2qf\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.904609 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.905995 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6" (UID: "1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.910515 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6" (UID: "1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:35 crc kubenswrapper[4815]: I0225 13:40:35.926135 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.007074 4815 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.007103 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.012197 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" event={"ID":"2e02f6a9-4025-45bc-a523-17b0c0758268","Type":"ContainerDied","Data":"8b65475db2ee56caa78e1c32b26aba990137f8fe1a5a4450720600f3f92735a2"} Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.012256 4815 scope.go:117] "RemoveContainer" containerID="f2077ba5aa6728327f15f64309d6fecfec0e0dd7553e042a47e119e5d1a5a998" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.012218 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-2wmhn" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.016960 4815 generic.go:334] "Generic (PLEG): container finished" podID="1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6" containerID="8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73" exitCode=0 Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.016981 4815 generic.go:334] "Generic (PLEG): container finished" podID="1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6" containerID="b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd" exitCode=143 Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.017019 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6","Type":"ContainerDied","Data":"8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73"} Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.017045 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6","Type":"ContainerDied","Data":"b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd"} Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.017055 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6","Type":"ContainerDied","Data":"498640fed436778f248eabe37882517d60bfae2afb5e5b6f8035d2abd386398c"} Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.017109 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.036430 4815 scope.go:117] "RemoveContainer" containerID="466bc2996fc37fe1a22e3361b2eeffbadf14ff5bfc43b0515011ba2561ad00c6" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.038263 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="20edc318-3cd1-42a8-ac54-728c466e8e4c" containerName="nova-api-log" containerID="cri-o://2012d1a4dc530731bf43543bc21dd983c7a3405bbd77e525c7e19653d98467ea" gracePeriod=30 Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.038411 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501","Type":"ContainerStarted","Data":"705c6d5290adfb067aacbde2a3678a6f0f684cf6dca17cfb57a3086c6452c308"} Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.038575 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="20edc318-3cd1-42a8-ac54-728c466e8e4c" containerName="nova-api-api" containerID="cri-o://c155b9b7d6d0ab409fbfb22530a8d29bf9183f4ca91ebc5cc0433c864eefabdd" gracePeriod=30 Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.074237 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2wmhn"] Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.077699 4815 scope.go:117] "RemoveContainer" containerID="8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.089598 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2wmhn"] Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.102189 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.128675 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.137788 4815 scope.go:117] "RemoveContainer" containerID="b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.145017 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:40:36 crc kubenswrapper[4815]: E0225 13:40:36.145531 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e02f6a9-4025-45bc-a523-17b0c0758268" containerName="dnsmasq-dns" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.145553 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e02f6a9-4025-45bc-a523-17b0c0758268" containerName="dnsmasq-dns" Feb 25 13:40:36 crc kubenswrapper[4815]: E0225 13:40:36.145585 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6" containerName="nova-metadata-log" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.145598 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6" containerName="nova-metadata-log" Feb 25 13:40:36 crc kubenswrapper[4815]: E0225 13:40:36.145615 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6" containerName="nova-metadata-metadata" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.145623 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6" containerName="nova-metadata-metadata" Feb 25 13:40:36 crc kubenswrapper[4815]: E0225 13:40:36.145649 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e02f6a9-4025-45bc-a523-17b0c0758268" containerName="init" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.145655 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e02f6a9-4025-45bc-a523-17b0c0758268" containerName="init" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.145826 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e02f6a9-4025-45bc-a523-17b0c0758268" containerName="dnsmasq-dns" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.145846 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6" containerName="nova-metadata-metadata" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.145860 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6" containerName="nova-metadata-log" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.147256 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.149554 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.151354 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.153024 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.156836 4815 scope.go:117] "RemoveContainer" containerID="8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73" Feb 25 13:40:36 crc kubenswrapper[4815]: E0225 13:40:36.157609 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73\": container with ID starting with 8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73 not found: ID does not exist" containerID="8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.157692 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73"} err="failed to get container status \"8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73\": rpc error: code = NotFound desc = could not find container \"8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73\": container with ID starting with 8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73 not found: ID does not exist" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.157717 4815 scope.go:117] "RemoveContainer" containerID="b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd" Feb 25 13:40:36 crc kubenswrapper[4815]: E0225 13:40:36.158841 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd\": container with ID starting with b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd not found: ID does not exist" containerID="b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.158881 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd"} err="failed to get container status \"b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd\": rpc error: code = NotFound desc = could not find container \"b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd\": container with ID starting with b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd not found: ID does not exist" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.158907 4815 scope.go:117] "RemoveContainer" containerID="8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.159335 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73"} err="failed to get container status \"8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73\": rpc error: code = NotFound desc = could not find container \"8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73\": container with ID starting with 8fe27e8854d6dcbdce5753c8bef69614f1ec408a9bda87a98aa2798ce8828a73 not found: ID does not exist" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.159428 4815 scope.go:117] "RemoveContainer" containerID="b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.159737 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd"} err="failed to get container status \"b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd\": rpc error: code = NotFound desc = could not find container \"b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd\": container with ID starting with b14c281222dd396abc97a78212dd56195184121825207dba557fdf6491e78ecd not found: ID does not exist" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.216867 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-config-data\") pod \"nova-metadata-0\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.217778 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.217898 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf5qf\" (UniqueName: \"kubernetes.io/projected/302dd34a-26ad-4c26-aa17-fe902863fd4b-kube-api-access-cf5qf\") pod \"nova-metadata-0\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.218004 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/302dd34a-26ad-4c26-aa17-fe902863fd4b-logs\") pod \"nova-metadata-0\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.218149 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.319923 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/302dd34a-26ad-4c26-aa17-fe902863fd4b-logs\") pod \"nova-metadata-0\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.320343 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.320417 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-config-data\") pod \"nova-metadata-0\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.320442 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/302dd34a-26ad-4c26-aa17-fe902863fd4b-logs\") pod \"nova-metadata-0\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.320459 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.320589 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf5qf\" (UniqueName: \"kubernetes.io/projected/302dd34a-26ad-4c26-aa17-fe902863fd4b-kube-api-access-cf5qf\") pod \"nova-metadata-0\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.328990 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.330066 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-config-data\") pod \"nova-metadata-0\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.339489 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.343023 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf5qf\" (UniqueName: \"kubernetes.io/projected/302dd34a-26ad-4c26-aa17-fe902863fd4b-kube-api-access-cf5qf\") pod \"nova-metadata-0\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.480586 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 13:40:36 crc kubenswrapper[4815]: W0225 13:40:36.945449 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod302dd34a_26ad_4c26_aa17_fe902863fd4b.slice/crio-11e35352b80a044ae4b0b6fadbb7f3c5b6d9e9e7dc5a08e2cb31a1f61fdac0a8 WatchSource:0}: Error finding container 11e35352b80a044ae4b0b6fadbb7f3c5b6d9e9e7dc5a08e2cb31a1f61fdac0a8: Status 404 returned error can't find the container with id 11e35352b80a044ae4b0b6fadbb7f3c5b6d9e9e7dc5a08e2cb31a1f61fdac0a8 Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.951927 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6" path="/var/lib/kubelet/pods/1e68dd8f-3901-4a33-b2a2-c3db0aef0ab6/volumes" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.952945 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e02f6a9-4025-45bc-a523-17b0c0758268" path="/var/lib/kubelet/pods/2e02f6a9-4025-45bc-a523-17b0c0758268/volumes" Feb 25 13:40:36 crc kubenswrapper[4815]: I0225 13:40:36.953797 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:40:37 crc kubenswrapper[4815]: I0225 13:40:37.057412 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501","Type":"ContainerStarted","Data":"143a0428700eeef8dfa8400407406dcc13caa20122248f9770b1a3646dcd581f"} Feb 25 13:40:37 crc kubenswrapper[4815]: I0225 13:40:37.057556 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 25 13:40:37 crc kubenswrapper[4815]: I0225 13:40:37.061638 4815 generic.go:334] "Generic (PLEG): container finished" podID="20edc318-3cd1-42a8-ac54-728c466e8e4c" containerID="2012d1a4dc530731bf43543bc21dd983c7a3405bbd77e525c7e19653d98467ea" exitCode=143 Feb 25 13:40:37 crc kubenswrapper[4815]: I0225 13:40:37.061754 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20edc318-3cd1-42a8-ac54-728c466e8e4c","Type":"ContainerDied","Data":"2012d1a4dc530731bf43543bc21dd983c7a3405bbd77e525c7e19653d98467ea"} Feb 25 13:40:37 crc kubenswrapper[4815]: I0225 13:40:37.063787 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"302dd34a-26ad-4c26-aa17-fe902863fd4b","Type":"ContainerStarted","Data":"11e35352b80a044ae4b0b6fadbb7f3c5b6d9e9e7dc5a08e2cb31a1f61fdac0a8"} Feb 25 13:40:37 crc kubenswrapper[4815]: I0225 13:40:37.065923 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="adaeacfe-023a-414e-a0c4-c59d8a181427" containerName="nova-scheduler-scheduler" containerID="cri-o://7178c331681a8ef2a0963b41b61126e06059b346fa5d5c111f6889cfb4923472" gracePeriod=30 Feb 25 13:40:37 crc kubenswrapper[4815]: I0225 13:40:37.090406 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.090366963 podStartE2EDuration="2.090366963s" podCreationTimestamp="2026-02-25 13:40:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:40:37.080647148 +0000 UTC m=+1194.881745202" watchObservedRunningTime="2026-02-25 13:40:37.090366963 +0000 UTC m=+1194.891465027" Feb 25 13:40:38 crc kubenswrapper[4815]: I0225 13:40:38.081571 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"302dd34a-26ad-4c26-aa17-fe902863fd4b","Type":"ContainerStarted","Data":"3d51ebbf50ce3f86ae89020d264c8ca488baa04f458010b1dd140ccca80f28a1"} Feb 25 13:40:38 crc kubenswrapper[4815]: I0225 13:40:38.082069 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"302dd34a-26ad-4c26-aa17-fe902863fd4b","Type":"ContainerStarted","Data":"65d029120089702d344365518d9dc68dc834a9c72765245c25c72ca180c5dd91"} Feb 25 13:40:38 crc kubenswrapper[4815]: I0225 13:40:38.111388 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.111368963 podStartE2EDuration="2.111368963s" podCreationTimestamp="2026-02-25 13:40:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:40:38.108144055 +0000 UTC m=+1195.909242149" watchObservedRunningTime="2026-02-25 13:40:38.111368963 +0000 UTC m=+1195.912467027" Feb 25 13:40:39 crc kubenswrapper[4815]: E0225 13:40:39.336331 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7178c331681a8ef2a0963b41b61126e06059b346fa5d5c111f6889cfb4923472" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 13:40:39 crc kubenswrapper[4815]: E0225 13:40:39.338212 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7178c331681a8ef2a0963b41b61126e06059b346fa5d5c111f6889cfb4923472" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 13:40:39 crc kubenswrapper[4815]: E0225 13:40:39.339604 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7178c331681a8ef2a0963b41b61126e06059b346fa5d5c111f6889cfb4923472" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 13:40:39 crc kubenswrapper[4815]: E0225 13:40:39.339777 4815 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="adaeacfe-023a-414e-a0c4-c59d8a181427" containerName="nova-scheduler-scheduler" Feb 25 13:40:40 crc kubenswrapper[4815]: I0225 13:40:40.987886 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.091208 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.114756 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20edc318-3cd1-42a8-ac54-728c466e8e4c-logs\") pod \"20edc318-3cd1-42a8-ac54-728c466e8e4c\" (UID: \"20edc318-3cd1-42a8-ac54-728c466e8e4c\") " Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.114823 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20edc318-3cd1-42a8-ac54-728c466e8e4c-combined-ca-bundle\") pod \"20edc318-3cd1-42a8-ac54-728c466e8e4c\" (UID: \"20edc318-3cd1-42a8-ac54-728c466e8e4c\") " Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.114891 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5j7sj\" (UniqueName: \"kubernetes.io/projected/20edc318-3cd1-42a8-ac54-728c466e8e4c-kube-api-access-5j7sj\") pod \"20edc318-3cd1-42a8-ac54-728c466e8e4c\" (UID: \"20edc318-3cd1-42a8-ac54-728c466e8e4c\") " Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.114977 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20edc318-3cd1-42a8-ac54-728c466e8e4c-config-data\") pod \"20edc318-3cd1-42a8-ac54-728c466e8e4c\" (UID: \"20edc318-3cd1-42a8-ac54-728c466e8e4c\") " Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.115309 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20edc318-3cd1-42a8-ac54-728c466e8e4c-logs" (OuterVolumeSpecName: "logs") pod "20edc318-3cd1-42a8-ac54-728c466e8e4c" (UID: "20edc318-3cd1-42a8-ac54-728c466e8e4c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.115675 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20edc318-3cd1-42a8-ac54-728c466e8e4c-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.122991 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20edc318-3cd1-42a8-ac54-728c466e8e4c-kube-api-access-5j7sj" (OuterVolumeSpecName: "kube-api-access-5j7sj") pod "20edc318-3cd1-42a8-ac54-728c466e8e4c" (UID: "20edc318-3cd1-42a8-ac54-728c466e8e4c"). InnerVolumeSpecName "kube-api-access-5j7sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.140328 4815 generic.go:334] "Generic (PLEG): container finished" podID="adaeacfe-023a-414e-a0c4-c59d8a181427" containerID="7178c331681a8ef2a0963b41b61126e06059b346fa5d5c111f6889cfb4923472" exitCode=0 Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.140431 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"adaeacfe-023a-414e-a0c4-c59d8a181427","Type":"ContainerDied","Data":"7178c331681a8ef2a0963b41b61126e06059b346fa5d5c111f6889cfb4923472"} Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.140459 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"adaeacfe-023a-414e-a0c4-c59d8a181427","Type":"ContainerDied","Data":"912c66e884cbc17db37b3073fc2bbeb101b6aff0d6e85e81e687c095e37d2700"} Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.140401 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.140474 4815 scope.go:117] "RemoveContainer" containerID="7178c331681a8ef2a0963b41b61126e06059b346fa5d5c111f6889cfb4923472" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.148226 4815 generic.go:334] "Generic (PLEG): container finished" podID="20edc318-3cd1-42a8-ac54-728c466e8e4c" containerID="c155b9b7d6d0ab409fbfb22530a8d29bf9183f4ca91ebc5cc0433c864eefabdd" exitCode=0 Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.148272 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20edc318-3cd1-42a8-ac54-728c466e8e4c","Type":"ContainerDied","Data":"c155b9b7d6d0ab409fbfb22530a8d29bf9183f4ca91ebc5cc0433c864eefabdd"} Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.148296 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20edc318-3cd1-42a8-ac54-728c466e8e4c","Type":"ContainerDied","Data":"f645b7292920975d94016d82de4c8de2586dffcf620c1d6a76fe5610f9b89494"} Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.148296 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.153618 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20edc318-3cd1-42a8-ac54-728c466e8e4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20edc318-3cd1-42a8-ac54-728c466e8e4c" (UID: "20edc318-3cd1-42a8-ac54-728c466e8e4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.167469 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20edc318-3cd1-42a8-ac54-728c466e8e4c-config-data" (OuterVolumeSpecName: "config-data") pod "20edc318-3cd1-42a8-ac54-728c466e8e4c" (UID: "20edc318-3cd1-42a8-ac54-728c466e8e4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.168635 4815 scope.go:117] "RemoveContainer" containerID="7178c331681a8ef2a0963b41b61126e06059b346fa5d5c111f6889cfb4923472" Feb 25 13:40:41 crc kubenswrapper[4815]: E0225 13:40:41.169135 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7178c331681a8ef2a0963b41b61126e06059b346fa5d5c111f6889cfb4923472\": container with ID starting with 7178c331681a8ef2a0963b41b61126e06059b346fa5d5c111f6889cfb4923472 not found: ID does not exist" containerID="7178c331681a8ef2a0963b41b61126e06059b346fa5d5c111f6889cfb4923472" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.169193 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7178c331681a8ef2a0963b41b61126e06059b346fa5d5c111f6889cfb4923472"} err="failed to get container status \"7178c331681a8ef2a0963b41b61126e06059b346fa5d5c111f6889cfb4923472\": rpc error: code = NotFound desc = could not find container \"7178c331681a8ef2a0963b41b61126e06059b346fa5d5c111f6889cfb4923472\": container with ID starting with 7178c331681a8ef2a0963b41b61126e06059b346fa5d5c111f6889cfb4923472 not found: ID does not exist" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.169226 4815 scope.go:117] "RemoveContainer" containerID="c155b9b7d6d0ab409fbfb22530a8d29bf9183f4ca91ebc5cc0433c864eefabdd" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.195019 4815 scope.go:117] "RemoveContainer" containerID="2012d1a4dc530731bf43543bc21dd983c7a3405bbd77e525c7e19653d98467ea" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.216300 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adaeacfe-023a-414e-a0c4-c59d8a181427-config-data\") pod \"adaeacfe-023a-414e-a0c4-c59d8a181427\" (UID: \"adaeacfe-023a-414e-a0c4-c59d8a181427\") " Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.216578 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqdgw\" (UniqueName: \"kubernetes.io/projected/adaeacfe-023a-414e-a0c4-c59d8a181427-kube-api-access-dqdgw\") pod \"adaeacfe-023a-414e-a0c4-c59d8a181427\" (UID: \"adaeacfe-023a-414e-a0c4-c59d8a181427\") " Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.216621 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adaeacfe-023a-414e-a0c4-c59d8a181427-combined-ca-bundle\") pod \"adaeacfe-023a-414e-a0c4-c59d8a181427\" (UID: \"adaeacfe-023a-414e-a0c4-c59d8a181427\") " Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.217543 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20edc318-3cd1-42a8-ac54-728c466e8e4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.217570 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5j7sj\" (UniqueName: \"kubernetes.io/projected/20edc318-3cd1-42a8-ac54-728c466e8e4c-kube-api-access-5j7sj\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.217585 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20edc318-3cd1-42a8-ac54-728c466e8e4c-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.220715 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adaeacfe-023a-414e-a0c4-c59d8a181427-kube-api-access-dqdgw" (OuterVolumeSpecName: "kube-api-access-dqdgw") pod "adaeacfe-023a-414e-a0c4-c59d8a181427" (UID: "adaeacfe-023a-414e-a0c4-c59d8a181427"). InnerVolumeSpecName "kube-api-access-dqdgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.228498 4815 scope.go:117] "RemoveContainer" containerID="c155b9b7d6d0ab409fbfb22530a8d29bf9183f4ca91ebc5cc0433c864eefabdd" Feb 25 13:40:41 crc kubenswrapper[4815]: E0225 13:40:41.229083 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c155b9b7d6d0ab409fbfb22530a8d29bf9183f4ca91ebc5cc0433c864eefabdd\": container with ID starting with c155b9b7d6d0ab409fbfb22530a8d29bf9183f4ca91ebc5cc0433c864eefabdd not found: ID does not exist" containerID="c155b9b7d6d0ab409fbfb22530a8d29bf9183f4ca91ebc5cc0433c864eefabdd" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.229124 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c155b9b7d6d0ab409fbfb22530a8d29bf9183f4ca91ebc5cc0433c864eefabdd"} err="failed to get container status \"c155b9b7d6d0ab409fbfb22530a8d29bf9183f4ca91ebc5cc0433c864eefabdd\": rpc error: code = NotFound desc = could not find container \"c155b9b7d6d0ab409fbfb22530a8d29bf9183f4ca91ebc5cc0433c864eefabdd\": container with ID starting with c155b9b7d6d0ab409fbfb22530a8d29bf9183f4ca91ebc5cc0433c864eefabdd not found: ID does not exist" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.229158 4815 scope.go:117] "RemoveContainer" containerID="2012d1a4dc530731bf43543bc21dd983c7a3405bbd77e525c7e19653d98467ea" Feb 25 13:40:41 crc kubenswrapper[4815]: E0225 13:40:41.229768 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2012d1a4dc530731bf43543bc21dd983c7a3405bbd77e525c7e19653d98467ea\": container with ID starting with 2012d1a4dc530731bf43543bc21dd983c7a3405bbd77e525c7e19653d98467ea not found: ID does not exist" containerID="2012d1a4dc530731bf43543bc21dd983c7a3405bbd77e525c7e19653d98467ea" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.229816 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2012d1a4dc530731bf43543bc21dd983c7a3405bbd77e525c7e19653d98467ea"} err="failed to get container status \"2012d1a4dc530731bf43543bc21dd983c7a3405bbd77e525c7e19653d98467ea\": rpc error: code = NotFound desc = could not find container \"2012d1a4dc530731bf43543bc21dd983c7a3405bbd77e525c7e19653d98467ea\": container with ID starting with 2012d1a4dc530731bf43543bc21dd983c7a3405bbd77e525c7e19653d98467ea not found: ID does not exist" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.238782 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adaeacfe-023a-414e-a0c4-c59d8a181427-config-data" (OuterVolumeSpecName: "config-data") pod "adaeacfe-023a-414e-a0c4-c59d8a181427" (UID: "adaeacfe-023a-414e-a0c4-c59d8a181427"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.262239 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adaeacfe-023a-414e-a0c4-c59d8a181427-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "adaeacfe-023a-414e-a0c4-c59d8a181427" (UID: "adaeacfe-023a-414e-a0c4-c59d8a181427"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.320615 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adaeacfe-023a-414e-a0c4-c59d8a181427-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.320649 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adaeacfe-023a-414e-a0c4-c59d8a181427-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.320661 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqdgw\" (UniqueName: \"kubernetes.io/projected/adaeacfe-023a-414e-a0c4-c59d8a181427-kube-api-access-dqdgw\") on node \"crc\" DevicePath \"\"" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.480950 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.481085 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.488186 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.502768 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.523834 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.543610 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.557108 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 13:40:41 crc kubenswrapper[4815]: E0225 13:40:41.557580 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20edc318-3cd1-42a8-ac54-728c466e8e4c" containerName="nova-api-log" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.557606 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="20edc318-3cd1-42a8-ac54-728c466e8e4c" containerName="nova-api-log" Feb 25 13:40:41 crc kubenswrapper[4815]: E0225 13:40:41.557626 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20edc318-3cd1-42a8-ac54-728c466e8e4c" containerName="nova-api-api" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.557634 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="20edc318-3cd1-42a8-ac54-728c466e8e4c" containerName="nova-api-api" Feb 25 13:40:41 crc kubenswrapper[4815]: E0225 13:40:41.557655 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adaeacfe-023a-414e-a0c4-c59d8a181427" containerName="nova-scheduler-scheduler" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.557663 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="adaeacfe-023a-414e-a0c4-c59d8a181427" containerName="nova-scheduler-scheduler" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.557868 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="20edc318-3cd1-42a8-ac54-728c466e8e4c" containerName="nova-api-api" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.557895 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="20edc318-3cd1-42a8-ac54-728c466e8e4c" containerName="nova-api-log" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.557912 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="adaeacfe-023a-414e-a0c4-c59d8a181427" containerName="nova-scheduler-scheduler" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.558644 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.561111 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.571354 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.582400 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.584376 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.588586 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.592875 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.625939 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86841bf5-3d74-480b-b7b4-ffb6534fb441-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"86841bf5-3d74-480b-b7b4-ffb6534fb441\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.626101 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmkmb\" (UniqueName: \"kubernetes.io/projected/86841bf5-3d74-480b-b7b4-ffb6534fb441-kube-api-access-bmkmb\") pod \"nova-scheduler-0\" (UID: \"86841bf5-3d74-480b-b7b4-ffb6534fb441\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.626150 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86841bf5-3d74-480b-b7b4-ffb6534fb441-config-data\") pod \"nova-scheduler-0\" (UID: \"86841bf5-3d74-480b-b7b4-ffb6534fb441\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.728754 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86841bf5-3d74-480b-b7b4-ffb6534fb441-config-data\") pod \"nova-scheduler-0\" (UID: \"86841bf5-3d74-480b-b7b4-ffb6534fb441\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.728850 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbllb\" (UniqueName: \"kubernetes.io/projected/0a22e2aa-a70a-4b98-add7-59a629664422-kube-api-access-lbllb\") pod \"nova-api-0\" (UID: \"0a22e2aa-a70a-4b98-add7-59a629664422\") " pod="openstack/nova-api-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.728883 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a22e2aa-a70a-4b98-add7-59a629664422-logs\") pod \"nova-api-0\" (UID: \"0a22e2aa-a70a-4b98-add7-59a629664422\") " pod="openstack/nova-api-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.729076 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a22e2aa-a70a-4b98-add7-59a629664422-config-data\") pod \"nova-api-0\" (UID: \"0a22e2aa-a70a-4b98-add7-59a629664422\") " pod="openstack/nova-api-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.729124 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86841bf5-3d74-480b-b7b4-ffb6534fb441-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"86841bf5-3d74-480b-b7b4-ffb6534fb441\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.729180 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a22e2aa-a70a-4b98-add7-59a629664422-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a22e2aa-a70a-4b98-add7-59a629664422\") " pod="openstack/nova-api-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.729261 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmkmb\" (UniqueName: \"kubernetes.io/projected/86841bf5-3d74-480b-b7b4-ffb6534fb441-kube-api-access-bmkmb\") pod \"nova-scheduler-0\" (UID: \"86841bf5-3d74-480b-b7b4-ffb6534fb441\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.731942 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86841bf5-3d74-480b-b7b4-ffb6534fb441-config-data\") pod \"nova-scheduler-0\" (UID: \"86841bf5-3d74-480b-b7b4-ffb6534fb441\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.739358 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86841bf5-3d74-480b-b7b4-ffb6534fb441-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"86841bf5-3d74-480b-b7b4-ffb6534fb441\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.749267 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmkmb\" (UniqueName: \"kubernetes.io/projected/86841bf5-3d74-480b-b7b4-ffb6534fb441-kube-api-access-bmkmb\") pod \"nova-scheduler-0\" (UID: \"86841bf5-3d74-480b-b7b4-ffb6534fb441\") " pod="openstack/nova-scheduler-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.831733 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbllb\" (UniqueName: \"kubernetes.io/projected/0a22e2aa-a70a-4b98-add7-59a629664422-kube-api-access-lbllb\") pod \"nova-api-0\" (UID: \"0a22e2aa-a70a-4b98-add7-59a629664422\") " pod="openstack/nova-api-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.832036 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a22e2aa-a70a-4b98-add7-59a629664422-logs\") pod \"nova-api-0\" (UID: \"0a22e2aa-a70a-4b98-add7-59a629664422\") " pod="openstack/nova-api-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.832262 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a22e2aa-a70a-4b98-add7-59a629664422-config-data\") pod \"nova-api-0\" (UID: \"0a22e2aa-a70a-4b98-add7-59a629664422\") " pod="openstack/nova-api-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.832454 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a22e2aa-a70a-4b98-add7-59a629664422-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a22e2aa-a70a-4b98-add7-59a629664422\") " pod="openstack/nova-api-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.832573 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a22e2aa-a70a-4b98-add7-59a629664422-logs\") pod \"nova-api-0\" (UID: \"0a22e2aa-a70a-4b98-add7-59a629664422\") " pod="openstack/nova-api-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.836363 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a22e2aa-a70a-4b98-add7-59a629664422-config-data\") pod \"nova-api-0\" (UID: \"0a22e2aa-a70a-4b98-add7-59a629664422\") " pod="openstack/nova-api-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.837996 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a22e2aa-a70a-4b98-add7-59a629664422-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a22e2aa-a70a-4b98-add7-59a629664422\") " pod="openstack/nova-api-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.853833 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbllb\" (UniqueName: \"kubernetes.io/projected/0a22e2aa-a70a-4b98-add7-59a629664422-kube-api-access-lbllb\") pod \"nova-api-0\" (UID: \"0a22e2aa-a70a-4b98-add7-59a629664422\") " pod="openstack/nova-api-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.894939 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 13:40:41 crc kubenswrapper[4815]: I0225 13:40:41.906171 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 13:40:42 crc kubenswrapper[4815]: W0225 13:40:42.378166 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86841bf5_3d74_480b_b7b4_ffb6534fb441.slice/crio-99e19d712ffce1dd44b94d8510b89b7b86676c84a2394e828ca790fecc44278b WatchSource:0}: Error finding container 99e19d712ffce1dd44b94d8510b89b7b86676c84a2394e828ca790fecc44278b: Status 404 returned error can't find the container with id 99e19d712ffce1dd44b94d8510b89b7b86676c84a2394e828ca790fecc44278b Feb 25 13:40:42 crc kubenswrapper[4815]: W0225 13:40:42.380024 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a22e2aa_a70a_4b98_add7_59a629664422.slice/crio-8a2dc94e37b560f7f69ddcf16065f1cf89c83821aba94d3dfaaff70c273cefa7 WatchSource:0}: Error finding container 8a2dc94e37b560f7f69ddcf16065f1cf89c83821aba94d3dfaaff70c273cefa7: Status 404 returned error can't find the container with id 8a2dc94e37b560f7f69ddcf16065f1cf89c83821aba94d3dfaaff70c273cefa7 Feb 25 13:40:42 crc kubenswrapper[4815]: I0225 13:40:42.383293 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 13:40:42 crc kubenswrapper[4815]: I0225 13:40:42.395350 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 13:40:42 crc kubenswrapper[4815]: I0225 13:40:42.964170 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20edc318-3cd1-42a8-ac54-728c466e8e4c" path="/var/lib/kubelet/pods/20edc318-3cd1-42a8-ac54-728c466e8e4c/volumes" Feb 25 13:40:42 crc kubenswrapper[4815]: I0225 13:40:42.966253 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adaeacfe-023a-414e-a0c4-c59d8a181427" path="/var/lib/kubelet/pods/adaeacfe-023a-414e-a0c4-c59d8a181427/volumes" Feb 25 13:40:43 crc kubenswrapper[4815]: I0225 13:40:43.179761 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a22e2aa-a70a-4b98-add7-59a629664422","Type":"ContainerStarted","Data":"529699ef5568c18880e63e007dbe80e7e9f62b8f607576a14fa709082ffdc09f"} Feb 25 13:40:43 crc kubenswrapper[4815]: I0225 13:40:43.180302 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a22e2aa-a70a-4b98-add7-59a629664422","Type":"ContainerStarted","Data":"b5f017ccd75437711dcd08527af92a987b91acad5265bec3e16df3b11227f5b2"} Feb 25 13:40:43 crc kubenswrapper[4815]: I0225 13:40:43.180344 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a22e2aa-a70a-4b98-add7-59a629664422","Type":"ContainerStarted","Data":"8a2dc94e37b560f7f69ddcf16065f1cf89c83821aba94d3dfaaff70c273cefa7"} Feb 25 13:40:43 crc kubenswrapper[4815]: I0225 13:40:43.186644 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"86841bf5-3d74-480b-b7b4-ffb6534fb441","Type":"ContainerStarted","Data":"f128963912d5f34faf0049947b1af6fc5870933479b5c5de0d1791f6da26f212"} Feb 25 13:40:43 crc kubenswrapper[4815]: I0225 13:40:43.186905 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"86841bf5-3d74-480b-b7b4-ffb6534fb441","Type":"ContainerStarted","Data":"99e19d712ffce1dd44b94d8510b89b7b86676c84a2394e828ca790fecc44278b"} Feb 25 13:40:43 crc kubenswrapper[4815]: I0225 13:40:43.205756 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.205737948 podStartE2EDuration="2.205737948s" podCreationTimestamp="2026-02-25 13:40:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:40:43.204894133 +0000 UTC m=+1201.005992227" watchObservedRunningTime="2026-02-25 13:40:43.205737948 +0000 UTC m=+1201.006836002" Feb 25 13:40:43 crc kubenswrapper[4815]: I0225 13:40:43.234728 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.234709106 podStartE2EDuration="2.234709106s" podCreationTimestamp="2026-02-25 13:40:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:40:43.233088177 +0000 UTC m=+1201.034186231" watchObservedRunningTime="2026-02-25 13:40:43.234709106 +0000 UTC m=+1201.035807160" Feb 25 13:40:44 crc kubenswrapper[4815]: I0225 13:40:44.880959 4815 scope.go:117] "RemoveContainer" containerID="bb4fedbfd8b9b49fa52cde59b50a74925e6f86d35b4e160a2acf3b4a245a3bb5" Feb 25 13:40:45 crc kubenswrapper[4815]: I0225 13:40:45.483024 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 25 13:40:46 crc kubenswrapper[4815]: I0225 13:40:46.481249 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 25 13:40:46 crc kubenswrapper[4815]: I0225 13:40:46.481627 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 25 13:40:46 crc kubenswrapper[4815]: I0225 13:40:46.895421 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 25 13:40:47 crc kubenswrapper[4815]: I0225 13:40:47.496767 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="302dd34a-26ad-4c26-aa17-fe902863fd4b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 13:40:47 crc kubenswrapper[4815]: I0225 13:40:47.496782 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="302dd34a-26ad-4c26-aa17-fe902863fd4b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 13:40:51 crc kubenswrapper[4815]: I0225 13:40:51.896197 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 25 13:40:51 crc kubenswrapper[4815]: I0225 13:40:51.907420 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 13:40:51 crc kubenswrapper[4815]: I0225 13:40:51.907503 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 13:40:51 crc kubenswrapper[4815]: I0225 13:40:51.928184 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 25 13:40:52 crc kubenswrapper[4815]: I0225 13:40:52.332158 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 25 13:40:52 crc kubenswrapper[4815]: I0225 13:40:52.989799 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0a22e2aa-a70a-4b98-add7-59a629664422" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.208:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 13:40:52 crc kubenswrapper[4815]: I0225 13:40:52.989798 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0a22e2aa-a70a-4b98-add7-59a629664422" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.208:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 13:40:53 crc kubenswrapper[4815]: I0225 13:40:53.053048 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 25 13:40:56 crc kubenswrapper[4815]: I0225 13:40:56.486635 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 25 13:40:56 crc kubenswrapper[4815]: I0225 13:40:56.487917 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 25 13:40:56 crc kubenswrapper[4815]: I0225 13:40:56.519889 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 25 13:40:57 crc kubenswrapper[4815]: I0225 13:40:57.352235 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.366615 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.376598 4815 generic.go:334] "Generic (PLEG): container finished" podID="fd374bbd-13c7-4b9c-91f9-6dfcce5cf518" containerID="8585f19dc562745bbf47de77c1c3d99029b63418cb239cffbf7aedaa87fb5fe1" exitCode=137 Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.376676 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518","Type":"ContainerDied","Data":"8585f19dc562745bbf47de77c1c3d99029b63418cb239cffbf7aedaa87fb5fe1"} Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.376970 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518","Type":"ContainerDied","Data":"09759eded458667cda35054112141f367d35475a6786b35cd280584efdfa155d"} Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.376991 4815 scope.go:117] "RemoveContainer" containerID="8585f19dc562745bbf47de77c1c3d99029b63418cb239cffbf7aedaa87fb5fe1" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.376688 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.419320 4815 scope.go:117] "RemoveContainer" containerID="8585f19dc562745bbf47de77c1c3d99029b63418cb239cffbf7aedaa87fb5fe1" Feb 25 13:41:00 crc kubenswrapper[4815]: E0225 13:41:00.419953 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8585f19dc562745bbf47de77c1c3d99029b63418cb239cffbf7aedaa87fb5fe1\": container with ID starting with 8585f19dc562745bbf47de77c1c3d99029b63418cb239cffbf7aedaa87fb5fe1 not found: ID does not exist" containerID="8585f19dc562745bbf47de77c1c3d99029b63418cb239cffbf7aedaa87fb5fe1" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.419998 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8585f19dc562745bbf47de77c1c3d99029b63418cb239cffbf7aedaa87fb5fe1"} err="failed to get container status \"8585f19dc562745bbf47de77c1c3d99029b63418cb239cffbf7aedaa87fb5fe1\": rpc error: code = NotFound desc = could not find container \"8585f19dc562745bbf47de77c1c3d99029b63418cb239cffbf7aedaa87fb5fe1\": container with ID starting with 8585f19dc562745bbf47de77c1c3d99029b63418cb239cffbf7aedaa87fb5fe1 not found: ID does not exist" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.422457 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7g966\" (UniqueName: \"kubernetes.io/projected/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-kube-api-access-7g966\") pod \"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518\" (UID: \"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518\") " Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.422757 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-combined-ca-bundle\") pod \"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518\" (UID: \"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518\") " Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.422847 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-config-data\") pod \"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518\" (UID: \"fd374bbd-13c7-4b9c-91f9-6dfcce5cf518\") " Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.428128 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-kube-api-access-7g966" (OuterVolumeSpecName: "kube-api-access-7g966") pod "fd374bbd-13c7-4b9c-91f9-6dfcce5cf518" (UID: "fd374bbd-13c7-4b9c-91f9-6dfcce5cf518"). InnerVolumeSpecName "kube-api-access-7g966". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.450930 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd374bbd-13c7-4b9c-91f9-6dfcce5cf518" (UID: "fd374bbd-13c7-4b9c-91f9-6dfcce5cf518"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.460385 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-config-data" (OuterVolumeSpecName: "config-data") pod "fd374bbd-13c7-4b9c-91f9-6dfcce5cf518" (UID: "fd374bbd-13c7-4b9c-91f9-6dfcce5cf518"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.525204 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.525252 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.525266 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7g966\" (UniqueName: \"kubernetes.io/projected/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518-kube-api-access-7g966\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.715494 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.724502 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.748324 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 13:41:00 crc kubenswrapper[4815]: E0225 13:41:00.748796 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd374bbd-13c7-4b9c-91f9-6dfcce5cf518" containerName="nova-cell1-novncproxy-novncproxy" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.748809 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd374bbd-13c7-4b9c-91f9-6dfcce5cf518" containerName="nova-cell1-novncproxy-novncproxy" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.748992 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd374bbd-13c7-4b9c-91f9-6dfcce5cf518" containerName="nova-cell1-novncproxy-novncproxy" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.749664 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.760775 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.760964 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.761049 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.772931 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.831167 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2597b11-1b7f-4f37-b553-f1385e830db1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2597b11-1b7f-4f37-b553-f1385e830db1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.831316 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rttj7\" (UniqueName: \"kubernetes.io/projected/f2597b11-1b7f-4f37-b553-f1385e830db1-kube-api-access-rttj7\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2597b11-1b7f-4f37-b553-f1385e830db1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.831365 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2597b11-1b7f-4f37-b553-f1385e830db1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2597b11-1b7f-4f37-b553-f1385e830db1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.831419 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2597b11-1b7f-4f37-b553-f1385e830db1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2597b11-1b7f-4f37-b553-f1385e830db1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.831594 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2597b11-1b7f-4f37-b553-f1385e830db1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2597b11-1b7f-4f37-b553-f1385e830db1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.932994 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2597b11-1b7f-4f37-b553-f1385e830db1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2597b11-1b7f-4f37-b553-f1385e830db1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.933061 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2597b11-1b7f-4f37-b553-f1385e830db1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2597b11-1b7f-4f37-b553-f1385e830db1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.933164 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2597b11-1b7f-4f37-b553-f1385e830db1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2597b11-1b7f-4f37-b553-f1385e830db1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.933225 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2597b11-1b7f-4f37-b553-f1385e830db1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2597b11-1b7f-4f37-b553-f1385e830db1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.933293 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rttj7\" (UniqueName: \"kubernetes.io/projected/f2597b11-1b7f-4f37-b553-f1385e830db1-kube-api-access-rttj7\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2597b11-1b7f-4f37-b553-f1385e830db1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.938964 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2597b11-1b7f-4f37-b553-f1385e830db1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2597b11-1b7f-4f37-b553-f1385e830db1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.940929 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2597b11-1b7f-4f37-b553-f1385e830db1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2597b11-1b7f-4f37-b553-f1385e830db1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.941271 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2597b11-1b7f-4f37-b553-f1385e830db1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2597b11-1b7f-4f37-b553-f1385e830db1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.948497 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2597b11-1b7f-4f37-b553-f1385e830db1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2597b11-1b7f-4f37-b553-f1385e830db1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.955026 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd374bbd-13c7-4b9c-91f9-6dfcce5cf518" path="/var/lib/kubelet/pods/fd374bbd-13c7-4b9c-91f9-6dfcce5cf518/volumes" Feb 25 13:41:00 crc kubenswrapper[4815]: I0225 13:41:00.965111 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rttj7\" (UniqueName: \"kubernetes.io/projected/f2597b11-1b7f-4f37-b553-f1385e830db1-kube-api-access-rttj7\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2597b11-1b7f-4f37-b553-f1385e830db1\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:01 crc kubenswrapper[4815]: I0225 13:41:01.083737 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:01 crc kubenswrapper[4815]: I0225 13:41:01.582199 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 13:41:01 crc kubenswrapper[4815]: I0225 13:41:01.912187 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 25 13:41:01 crc kubenswrapper[4815]: I0225 13:41:01.913397 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 25 13:41:01 crc kubenswrapper[4815]: I0225 13:41:01.914226 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 25 13:41:01 crc kubenswrapper[4815]: I0225 13:41:01.922052 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.399824 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f2597b11-1b7f-4f37-b553-f1385e830db1","Type":"ContainerStarted","Data":"f7ca21e87837f245fd365f2e23e83e57d6594de56cf0db608df08b9867aae6c0"} Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.399875 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f2597b11-1b7f-4f37-b553-f1385e830db1","Type":"ContainerStarted","Data":"ae56d7e3c83498d6283a7fcfa57202944637d9d86119577b4d5f3261951b5912"} Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.400092 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.404236 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.420732 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.420714313 podStartE2EDuration="2.420714313s" podCreationTimestamp="2026-02-25 13:41:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:41:02.416150445 +0000 UTC m=+1220.217248509" watchObservedRunningTime="2026-02-25 13:41:02.420714313 +0000 UTC m=+1220.221812367" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.591645 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-mzfxs"] Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.610012 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.672182 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.672320 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.672358 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-config\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.672385 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dskch\" (UniqueName: \"kubernetes.io/projected/292520ae-3dbf-4627-bf1d-b815d5af271b-kube-api-access-dskch\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.672416 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.672435 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.691145 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-mzfxs"] Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.775305 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.775626 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-config\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.775742 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dskch\" (UniqueName: \"kubernetes.io/projected/292520ae-3dbf-4627-bf1d-b815d5af271b-kube-api-access-dskch\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.775830 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.775906 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.776025 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.776977 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.777553 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.778124 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-config\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.779210 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.779747 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:02 crc kubenswrapper[4815]: I0225 13:41:02.809189 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dskch\" (UniqueName: \"kubernetes.io/projected/292520ae-3dbf-4627-bf1d-b815d5af271b-kube-api-access-dskch\") pod \"dnsmasq-dns-5c7b6c5df9-mzfxs\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:03 crc kubenswrapper[4815]: I0225 13:41:03.013929 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:03 crc kubenswrapper[4815]: I0225 13:41:03.503298 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-mzfxs"] Feb 25 13:41:03 crc kubenswrapper[4815]: W0225 13:41:03.503836 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod292520ae_3dbf_4627_bf1d_b815d5af271b.slice/crio-38af34e7109228ad9e0196a315f45e433a0b2af650168afb4240537996b210fd WatchSource:0}: Error finding container 38af34e7109228ad9e0196a315f45e433a0b2af650168afb4240537996b210fd: Status 404 returned error can't find the container with id 38af34e7109228ad9e0196a315f45e433a0b2af650168afb4240537996b210fd Feb 25 13:41:04 crc kubenswrapper[4815]: I0225 13:41:04.417816 4815 generic.go:334] "Generic (PLEG): container finished" podID="292520ae-3dbf-4627-bf1d-b815d5af271b" containerID="37e0b33154bb461902a98e3c518566038390dafbcafab1ed57ea6babbc26aa7f" exitCode=0 Feb 25 13:41:04 crc kubenswrapper[4815]: I0225 13:41:04.417869 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" event={"ID":"292520ae-3dbf-4627-bf1d-b815d5af271b","Type":"ContainerDied","Data":"37e0b33154bb461902a98e3c518566038390dafbcafab1ed57ea6babbc26aa7f"} Feb 25 13:41:04 crc kubenswrapper[4815]: I0225 13:41:04.418165 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" event={"ID":"292520ae-3dbf-4627-bf1d-b815d5af271b","Type":"ContainerStarted","Data":"38af34e7109228ad9e0196a315f45e433a0b2af650168afb4240537996b210fd"} Feb 25 13:41:04 crc kubenswrapper[4815]: I0225 13:41:04.582407 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:41:04 crc kubenswrapper[4815]: I0225 13:41:04.582760 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b019b447-6312-4672-861b-2692658b739d" containerName="proxy-httpd" containerID="cri-o://8950b7b5ab3b75b777743934b8423157e85e74526e0934fe3a56d59911cc25d8" gracePeriod=30 Feb 25 13:41:04 crc kubenswrapper[4815]: I0225 13:41:04.583093 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b019b447-6312-4672-861b-2692658b739d" containerName="sg-core" containerID="cri-o://9dcbabae071a33cf72605cd940a5e2af845afd5a2f96afa808607ec367d5037a" gracePeriod=30 Feb 25 13:41:04 crc kubenswrapper[4815]: I0225 13:41:04.583165 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b019b447-6312-4672-861b-2692658b739d" containerName="ceilometer-central-agent" containerID="cri-o://d5e81f0eb1705fa7f0ce74497c1f61a949c0f6d42cbad1965f8c1e08a1ad7440" gracePeriod=30 Feb 25 13:41:04 crc kubenswrapper[4815]: I0225 13:41:04.583159 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b019b447-6312-4672-861b-2692658b739d" containerName="ceilometer-notification-agent" containerID="cri-o://5c6a67efff1d2c98e92ccb3c3727b0819482bb0d62c7f86fe25d2b1431ec54fa" gracePeriod=30 Feb 25 13:41:04 crc kubenswrapper[4815]: I0225 13:41:04.724657 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 13:41:05 crc kubenswrapper[4815]: I0225 13:41:05.427069 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" event={"ID":"292520ae-3dbf-4627-bf1d-b815d5af271b","Type":"ContainerStarted","Data":"3f4ff5bb3070ffc987474d20227164b2c607635014c7ac60afc26ad6d0d677ea"} Feb 25 13:41:05 crc kubenswrapper[4815]: I0225 13:41:05.427187 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:05 crc kubenswrapper[4815]: I0225 13:41:05.429582 4815 generic.go:334] "Generic (PLEG): container finished" podID="b019b447-6312-4672-861b-2692658b739d" containerID="8950b7b5ab3b75b777743934b8423157e85e74526e0934fe3a56d59911cc25d8" exitCode=0 Feb 25 13:41:05 crc kubenswrapper[4815]: I0225 13:41:05.429615 4815 generic.go:334] "Generic (PLEG): container finished" podID="b019b447-6312-4672-861b-2692658b739d" containerID="9dcbabae071a33cf72605cd940a5e2af845afd5a2f96afa808607ec367d5037a" exitCode=2 Feb 25 13:41:05 crc kubenswrapper[4815]: I0225 13:41:05.429628 4815 generic.go:334] "Generic (PLEG): container finished" podID="b019b447-6312-4672-861b-2692658b739d" containerID="d5e81f0eb1705fa7f0ce74497c1f61a949c0f6d42cbad1965f8c1e08a1ad7440" exitCode=0 Feb 25 13:41:05 crc kubenswrapper[4815]: I0225 13:41:05.429634 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b019b447-6312-4672-861b-2692658b739d","Type":"ContainerDied","Data":"8950b7b5ab3b75b777743934b8423157e85e74526e0934fe3a56d59911cc25d8"} Feb 25 13:41:05 crc kubenswrapper[4815]: I0225 13:41:05.429693 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b019b447-6312-4672-861b-2692658b739d","Type":"ContainerDied","Data":"9dcbabae071a33cf72605cd940a5e2af845afd5a2f96afa808607ec367d5037a"} Feb 25 13:41:05 crc kubenswrapper[4815]: I0225 13:41:05.429707 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b019b447-6312-4672-861b-2692658b739d","Type":"ContainerDied","Data":"d5e81f0eb1705fa7f0ce74497c1f61a949c0f6d42cbad1965f8c1e08a1ad7440"} Feb 25 13:41:05 crc kubenswrapper[4815]: I0225 13:41:05.429790 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0a22e2aa-a70a-4b98-add7-59a629664422" containerName="nova-api-log" containerID="cri-o://b5f017ccd75437711dcd08527af92a987b91acad5265bec3e16df3b11227f5b2" gracePeriod=30 Feb 25 13:41:05 crc kubenswrapper[4815]: I0225 13:41:05.429873 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0a22e2aa-a70a-4b98-add7-59a629664422" containerName="nova-api-api" containerID="cri-o://529699ef5568c18880e63e007dbe80e7e9f62b8f607576a14fa709082ffdc09f" gracePeriod=30 Feb 25 13:41:05 crc kubenswrapper[4815]: I0225 13:41:05.467153 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" podStartSLOduration=3.467127089 podStartE2EDuration="3.467127089s" podCreationTimestamp="2026-02-25 13:41:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:41:05.455379833 +0000 UTC m=+1223.256477897" watchObservedRunningTime="2026-02-25 13:41:05.467127089 +0000 UTC m=+1223.268225143" Feb 25 13:41:06 crc kubenswrapper[4815]: I0225 13:41:06.084854 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:06 crc kubenswrapper[4815]: I0225 13:41:06.440016 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a22e2aa-a70a-4b98-add7-59a629664422","Type":"ContainerDied","Data":"b5f017ccd75437711dcd08527af92a987b91acad5265bec3e16df3b11227f5b2"} Feb 25 13:41:06 crc kubenswrapper[4815]: I0225 13:41:06.439943 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a22e2aa-a70a-4b98-add7-59a629664422" containerID="b5f017ccd75437711dcd08527af92a987b91acad5265bec3e16df3b11227f5b2" exitCode=143 Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.419899 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.476387 4815 generic.go:334] "Generic (PLEG): container finished" podID="b019b447-6312-4672-861b-2692658b739d" containerID="5c6a67efff1d2c98e92ccb3c3727b0819482bb0d62c7f86fe25d2b1431ec54fa" exitCode=0 Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.476438 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b019b447-6312-4672-861b-2692658b739d","Type":"ContainerDied","Data":"5c6a67efff1d2c98e92ccb3c3727b0819482bb0d62c7f86fe25d2b1431ec54fa"} Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.476469 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b019b447-6312-4672-861b-2692658b739d","Type":"ContainerDied","Data":"d8ca2a8d15e3cb90df3ebc5f1f20223e16f77f79499f8f34e7321381e58851ce"} Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.476490 4815 scope.go:117] "RemoveContainer" containerID="8950b7b5ab3b75b777743934b8423157e85e74526e0934fe3a56d59911cc25d8" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.476848 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.481701 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gm9xh\" (UniqueName: \"kubernetes.io/projected/b019b447-6312-4672-861b-2692658b739d-kube-api-access-gm9xh\") pod \"b019b447-6312-4672-861b-2692658b739d\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.484492 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-scripts\") pod \"b019b447-6312-4672-861b-2692658b739d\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.485627 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b019b447-6312-4672-861b-2692658b739d-log-httpd\") pod \"b019b447-6312-4672-861b-2692658b739d\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.485776 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-config-data\") pod \"b019b447-6312-4672-861b-2692658b739d\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.485944 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-sg-core-conf-yaml\") pod \"b019b447-6312-4672-861b-2692658b739d\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.486127 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-ceilometer-tls-certs\") pod \"b019b447-6312-4672-861b-2692658b739d\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.486274 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b019b447-6312-4672-861b-2692658b739d-run-httpd\") pod \"b019b447-6312-4672-861b-2692658b739d\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.486408 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-combined-ca-bundle\") pod \"b019b447-6312-4672-861b-2692658b739d\" (UID: \"b019b447-6312-4672-861b-2692658b739d\") " Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.488878 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b019b447-6312-4672-861b-2692658b739d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b019b447-6312-4672-861b-2692658b739d" (UID: "b019b447-6312-4672-861b-2692658b739d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.488929 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b019b447-6312-4672-861b-2692658b739d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b019b447-6312-4672-861b-2692658b739d" (UID: "b019b447-6312-4672-861b-2692658b739d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.493130 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-scripts" (OuterVolumeSpecName: "scripts") pod "b019b447-6312-4672-861b-2692658b739d" (UID: "b019b447-6312-4672-861b-2692658b739d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.511055 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b019b447-6312-4672-861b-2692658b739d-kube-api-access-gm9xh" (OuterVolumeSpecName: "kube-api-access-gm9xh") pod "b019b447-6312-4672-861b-2692658b739d" (UID: "b019b447-6312-4672-861b-2692658b739d"). InnerVolumeSpecName "kube-api-access-gm9xh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.518735 4815 scope.go:117] "RemoveContainer" containerID="9dcbabae071a33cf72605cd940a5e2af845afd5a2f96afa808607ec367d5037a" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.529703 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b019b447-6312-4672-861b-2692658b739d" (UID: "b019b447-6312-4672-861b-2692658b739d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.537838 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "b019b447-6312-4672-861b-2692658b739d" (UID: "b019b447-6312-4672-861b-2692658b739d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.568285 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b019b447-6312-4672-861b-2692658b739d" (UID: "b019b447-6312-4672-861b-2692658b739d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.590151 4815 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.590214 4815 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b019b447-6312-4672-861b-2692658b739d-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.590235 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.590255 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gm9xh\" (UniqueName: \"kubernetes.io/projected/b019b447-6312-4672-861b-2692658b739d-kube-api-access-gm9xh\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.590273 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.590290 4815 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b019b447-6312-4672-861b-2692658b739d-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.590306 4815 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.606080 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-config-data" (OuterVolumeSpecName: "config-data") pod "b019b447-6312-4672-861b-2692658b739d" (UID: "b019b447-6312-4672-861b-2692658b739d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.670831 4815 scope.go:117] "RemoveContainer" containerID="5c6a67efff1d2c98e92ccb3c3727b0819482bb0d62c7f86fe25d2b1431ec54fa" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.687734 4815 scope.go:117] "RemoveContainer" containerID="d5e81f0eb1705fa7f0ce74497c1f61a949c0f6d42cbad1965f8c1e08a1ad7440" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.691598 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b019b447-6312-4672-861b-2692658b739d-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.708495 4815 scope.go:117] "RemoveContainer" containerID="8950b7b5ab3b75b777743934b8423157e85e74526e0934fe3a56d59911cc25d8" Feb 25 13:41:08 crc kubenswrapper[4815]: E0225 13:41:08.710156 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8950b7b5ab3b75b777743934b8423157e85e74526e0934fe3a56d59911cc25d8\": container with ID starting with 8950b7b5ab3b75b777743934b8423157e85e74526e0934fe3a56d59911cc25d8 not found: ID does not exist" containerID="8950b7b5ab3b75b777743934b8423157e85e74526e0934fe3a56d59911cc25d8" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.710196 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8950b7b5ab3b75b777743934b8423157e85e74526e0934fe3a56d59911cc25d8"} err="failed to get container status \"8950b7b5ab3b75b777743934b8423157e85e74526e0934fe3a56d59911cc25d8\": rpc error: code = NotFound desc = could not find container \"8950b7b5ab3b75b777743934b8423157e85e74526e0934fe3a56d59911cc25d8\": container with ID starting with 8950b7b5ab3b75b777743934b8423157e85e74526e0934fe3a56d59911cc25d8 not found: ID does not exist" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.710224 4815 scope.go:117] "RemoveContainer" containerID="9dcbabae071a33cf72605cd940a5e2af845afd5a2f96afa808607ec367d5037a" Feb 25 13:41:08 crc kubenswrapper[4815]: E0225 13:41:08.710538 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dcbabae071a33cf72605cd940a5e2af845afd5a2f96afa808607ec367d5037a\": container with ID starting with 9dcbabae071a33cf72605cd940a5e2af845afd5a2f96afa808607ec367d5037a not found: ID does not exist" containerID="9dcbabae071a33cf72605cd940a5e2af845afd5a2f96afa808607ec367d5037a" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.710565 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dcbabae071a33cf72605cd940a5e2af845afd5a2f96afa808607ec367d5037a"} err="failed to get container status \"9dcbabae071a33cf72605cd940a5e2af845afd5a2f96afa808607ec367d5037a\": rpc error: code = NotFound desc = could not find container \"9dcbabae071a33cf72605cd940a5e2af845afd5a2f96afa808607ec367d5037a\": container with ID starting with 9dcbabae071a33cf72605cd940a5e2af845afd5a2f96afa808607ec367d5037a not found: ID does not exist" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.710578 4815 scope.go:117] "RemoveContainer" containerID="5c6a67efff1d2c98e92ccb3c3727b0819482bb0d62c7f86fe25d2b1431ec54fa" Feb 25 13:41:08 crc kubenswrapper[4815]: E0225 13:41:08.710840 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c6a67efff1d2c98e92ccb3c3727b0819482bb0d62c7f86fe25d2b1431ec54fa\": container with ID starting with 5c6a67efff1d2c98e92ccb3c3727b0819482bb0d62c7f86fe25d2b1431ec54fa not found: ID does not exist" containerID="5c6a67efff1d2c98e92ccb3c3727b0819482bb0d62c7f86fe25d2b1431ec54fa" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.710877 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c6a67efff1d2c98e92ccb3c3727b0819482bb0d62c7f86fe25d2b1431ec54fa"} err="failed to get container status \"5c6a67efff1d2c98e92ccb3c3727b0819482bb0d62c7f86fe25d2b1431ec54fa\": rpc error: code = NotFound desc = could not find container \"5c6a67efff1d2c98e92ccb3c3727b0819482bb0d62c7f86fe25d2b1431ec54fa\": container with ID starting with 5c6a67efff1d2c98e92ccb3c3727b0819482bb0d62c7f86fe25d2b1431ec54fa not found: ID does not exist" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.710904 4815 scope.go:117] "RemoveContainer" containerID="d5e81f0eb1705fa7f0ce74497c1f61a949c0f6d42cbad1965f8c1e08a1ad7440" Feb 25 13:41:08 crc kubenswrapper[4815]: E0225 13:41:08.711316 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5e81f0eb1705fa7f0ce74497c1f61a949c0f6d42cbad1965f8c1e08a1ad7440\": container with ID starting with d5e81f0eb1705fa7f0ce74497c1f61a949c0f6d42cbad1965f8c1e08a1ad7440 not found: ID does not exist" containerID="d5e81f0eb1705fa7f0ce74497c1f61a949c0f6d42cbad1965f8c1e08a1ad7440" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.711345 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5e81f0eb1705fa7f0ce74497c1f61a949c0f6d42cbad1965f8c1e08a1ad7440"} err="failed to get container status \"d5e81f0eb1705fa7f0ce74497c1f61a949c0f6d42cbad1965f8c1e08a1ad7440\": rpc error: code = NotFound desc = could not find container \"d5e81f0eb1705fa7f0ce74497c1f61a949c0f6d42cbad1965f8c1e08a1ad7440\": container with ID starting with d5e81f0eb1705fa7f0ce74497c1f61a949c0f6d42cbad1965f8c1e08a1ad7440 not found: ID does not exist" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.822651 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.836275 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.850044 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:41:08 crc kubenswrapper[4815]: E0225 13:41:08.852500 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b019b447-6312-4672-861b-2692658b739d" containerName="ceilometer-central-agent" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.852580 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b019b447-6312-4672-861b-2692658b739d" containerName="ceilometer-central-agent" Feb 25 13:41:08 crc kubenswrapper[4815]: E0225 13:41:08.852652 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b019b447-6312-4672-861b-2692658b739d" containerName="sg-core" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.852661 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b019b447-6312-4672-861b-2692658b739d" containerName="sg-core" Feb 25 13:41:08 crc kubenswrapper[4815]: E0225 13:41:08.852730 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b019b447-6312-4672-861b-2692658b739d" containerName="proxy-httpd" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.852742 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b019b447-6312-4672-861b-2692658b739d" containerName="proxy-httpd" Feb 25 13:41:08 crc kubenswrapper[4815]: E0225 13:41:08.852913 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b019b447-6312-4672-861b-2692658b739d" containerName="ceilometer-notification-agent" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.852925 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b019b447-6312-4672-861b-2692658b739d" containerName="ceilometer-notification-agent" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.853542 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b019b447-6312-4672-861b-2692658b739d" containerName="sg-core" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.853569 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b019b447-6312-4672-861b-2692658b739d" containerName="proxy-httpd" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.853581 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b019b447-6312-4672-861b-2692658b739d" containerName="ceilometer-notification-agent" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.853597 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b019b447-6312-4672-861b-2692658b739d" containerName="ceilometer-central-agent" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.856327 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.859615 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.859831 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.860434 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.863492 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.895103 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98c6009e-1597-4edc-a7d9-3987e946bd57-run-httpd\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.895146 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/98c6009e-1597-4edc-a7d9-3987e946bd57-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.895188 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98c6009e-1597-4edc-a7d9-3987e946bd57-scripts\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.895266 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98c6009e-1597-4edc-a7d9-3987e946bd57-config-data\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.895281 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c6009e-1597-4edc-a7d9-3987e946bd57-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.895374 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffk25\" (UniqueName: \"kubernetes.io/projected/98c6009e-1597-4edc-a7d9-3987e946bd57-kube-api-access-ffk25\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.895398 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98c6009e-1597-4edc-a7d9-3987e946bd57-log-httpd\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.895430 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98c6009e-1597-4edc-a7d9-3987e946bd57-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.927468 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.952147 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b019b447-6312-4672-861b-2692658b739d" path="/var/lib/kubelet/pods/b019b447-6312-4672-861b-2692658b739d/volumes" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.997101 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbllb\" (UniqueName: \"kubernetes.io/projected/0a22e2aa-a70a-4b98-add7-59a629664422-kube-api-access-lbllb\") pod \"0a22e2aa-a70a-4b98-add7-59a629664422\" (UID: \"0a22e2aa-a70a-4b98-add7-59a629664422\") " Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.997563 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a22e2aa-a70a-4b98-add7-59a629664422-config-data\") pod \"0a22e2aa-a70a-4b98-add7-59a629664422\" (UID: \"0a22e2aa-a70a-4b98-add7-59a629664422\") " Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.997594 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a22e2aa-a70a-4b98-add7-59a629664422-combined-ca-bundle\") pod \"0a22e2aa-a70a-4b98-add7-59a629664422\" (UID: \"0a22e2aa-a70a-4b98-add7-59a629664422\") " Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.997860 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a22e2aa-a70a-4b98-add7-59a629664422-logs\") pod \"0a22e2aa-a70a-4b98-add7-59a629664422\" (UID: \"0a22e2aa-a70a-4b98-add7-59a629664422\") " Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.998113 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98c6009e-1597-4edc-a7d9-3987e946bd57-config-data\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.998142 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c6009e-1597-4edc-a7d9-3987e946bd57-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.998311 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffk25\" (UniqueName: \"kubernetes.io/projected/98c6009e-1597-4edc-a7d9-3987e946bd57-kube-api-access-ffk25\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.998355 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98c6009e-1597-4edc-a7d9-3987e946bd57-log-httpd\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.998403 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98c6009e-1597-4edc-a7d9-3987e946bd57-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.998458 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98c6009e-1597-4edc-a7d9-3987e946bd57-run-httpd\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.998485 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/98c6009e-1597-4edc-a7d9-3987e946bd57-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:08 crc kubenswrapper[4815]: I0225 13:41:08.998557 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98c6009e-1597-4edc-a7d9-3987e946bd57-scripts\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.002837 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a22e2aa-a70a-4b98-add7-59a629664422-logs" (OuterVolumeSpecName: "logs") pod "0a22e2aa-a70a-4b98-add7-59a629664422" (UID: "0a22e2aa-a70a-4b98-add7-59a629664422"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.003381 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98c6009e-1597-4edc-a7d9-3987e946bd57-log-httpd\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.003499 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98c6009e-1597-4edc-a7d9-3987e946bd57-run-httpd\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.003965 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98c6009e-1597-4edc-a7d9-3987e946bd57-scripts\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.006781 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98c6009e-1597-4edc-a7d9-3987e946bd57-config-data\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.007272 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c6009e-1597-4edc-a7d9-3987e946bd57-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.007704 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/98c6009e-1597-4edc-a7d9-3987e946bd57-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.008249 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98c6009e-1597-4edc-a7d9-3987e946bd57-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.018800 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a22e2aa-a70a-4b98-add7-59a629664422-kube-api-access-lbllb" (OuterVolumeSpecName: "kube-api-access-lbllb") pod "0a22e2aa-a70a-4b98-add7-59a629664422" (UID: "0a22e2aa-a70a-4b98-add7-59a629664422"). InnerVolumeSpecName "kube-api-access-lbllb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.037838 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffk25\" (UniqueName: \"kubernetes.io/projected/98c6009e-1597-4edc-a7d9-3987e946bd57-kube-api-access-ffk25\") pod \"ceilometer-0\" (UID: \"98c6009e-1597-4edc-a7d9-3987e946bd57\") " pod="openstack/ceilometer-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.044168 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a22e2aa-a70a-4b98-add7-59a629664422-config-data" (OuterVolumeSpecName: "config-data") pod "0a22e2aa-a70a-4b98-add7-59a629664422" (UID: "0a22e2aa-a70a-4b98-add7-59a629664422"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.046653 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a22e2aa-a70a-4b98-add7-59a629664422-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a22e2aa-a70a-4b98-add7-59a629664422" (UID: "0a22e2aa-a70a-4b98-add7-59a629664422"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.100587 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a22e2aa-a70a-4b98-add7-59a629664422-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.100632 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbllb\" (UniqueName: \"kubernetes.io/projected/0a22e2aa-a70a-4b98-add7-59a629664422-kube-api-access-lbllb\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.100645 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a22e2aa-a70a-4b98-add7-59a629664422-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.100656 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a22e2aa-a70a-4b98-add7-59a629664422-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.177574 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.534994 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a22e2aa-a70a-4b98-add7-59a629664422" containerID="529699ef5568c18880e63e007dbe80e7e9f62b8f607576a14fa709082ffdc09f" exitCode=0 Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.535103 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a22e2aa-a70a-4b98-add7-59a629664422","Type":"ContainerDied","Data":"529699ef5568c18880e63e007dbe80e7e9f62b8f607576a14fa709082ffdc09f"} Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.535300 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a22e2aa-a70a-4b98-add7-59a629664422","Type":"ContainerDied","Data":"8a2dc94e37b560f7f69ddcf16065f1cf89c83821aba94d3dfaaff70c273cefa7"} Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.535317 4815 scope.go:117] "RemoveContainer" containerID="529699ef5568c18880e63e007dbe80e7e9f62b8f607576a14fa709082ffdc09f" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.535164 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.560172 4815 scope.go:117] "RemoveContainer" containerID="b5f017ccd75437711dcd08527af92a987b91acad5265bec3e16df3b11227f5b2" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.576144 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.598936 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.599033 4815 scope.go:117] "RemoveContainer" containerID="529699ef5568c18880e63e007dbe80e7e9f62b8f607576a14fa709082ffdc09f" Feb 25 13:41:09 crc kubenswrapper[4815]: E0225 13:41:09.601965 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"529699ef5568c18880e63e007dbe80e7e9f62b8f607576a14fa709082ffdc09f\": container with ID starting with 529699ef5568c18880e63e007dbe80e7e9f62b8f607576a14fa709082ffdc09f not found: ID does not exist" containerID="529699ef5568c18880e63e007dbe80e7e9f62b8f607576a14fa709082ffdc09f" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.602007 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"529699ef5568c18880e63e007dbe80e7e9f62b8f607576a14fa709082ffdc09f"} err="failed to get container status \"529699ef5568c18880e63e007dbe80e7e9f62b8f607576a14fa709082ffdc09f\": rpc error: code = NotFound desc = could not find container \"529699ef5568c18880e63e007dbe80e7e9f62b8f607576a14fa709082ffdc09f\": container with ID starting with 529699ef5568c18880e63e007dbe80e7e9f62b8f607576a14fa709082ffdc09f not found: ID does not exist" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.602039 4815 scope.go:117] "RemoveContainer" containerID="b5f017ccd75437711dcd08527af92a987b91acad5265bec3e16df3b11227f5b2" Feb 25 13:41:09 crc kubenswrapper[4815]: E0225 13:41:09.602497 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5f017ccd75437711dcd08527af92a987b91acad5265bec3e16df3b11227f5b2\": container with ID starting with b5f017ccd75437711dcd08527af92a987b91acad5265bec3e16df3b11227f5b2 not found: ID does not exist" containerID="b5f017ccd75437711dcd08527af92a987b91acad5265bec3e16df3b11227f5b2" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.602547 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5f017ccd75437711dcd08527af92a987b91acad5265bec3e16df3b11227f5b2"} err="failed to get container status \"b5f017ccd75437711dcd08527af92a987b91acad5265bec3e16df3b11227f5b2\": rpc error: code = NotFound desc = could not find container \"b5f017ccd75437711dcd08527af92a987b91acad5265bec3e16df3b11227f5b2\": container with ID starting with b5f017ccd75437711dcd08527af92a987b91acad5265bec3e16df3b11227f5b2 not found: ID does not exist" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.616338 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 25 13:41:09 crc kubenswrapper[4815]: E0225 13:41:09.616856 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a22e2aa-a70a-4b98-add7-59a629664422" containerName="nova-api-api" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.616874 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a22e2aa-a70a-4b98-add7-59a629664422" containerName="nova-api-api" Feb 25 13:41:09 crc kubenswrapper[4815]: E0225 13:41:09.616888 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a22e2aa-a70a-4b98-add7-59a629664422" containerName="nova-api-log" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.616895 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a22e2aa-a70a-4b98-add7-59a629664422" containerName="nova-api-log" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.617121 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a22e2aa-a70a-4b98-add7-59a629664422" containerName="nova-api-api" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.617139 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a22e2aa-a70a-4b98-add7-59a629664422" containerName="nova-api-log" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.622962 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.625189 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.625386 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.625817 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.626332 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.646034 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.710793 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-config-data\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.710857 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-logs\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.710917 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ntjr\" (UniqueName: \"kubernetes.io/projected/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-kube-api-access-8ntjr\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.710947 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.710978 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.710993 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-public-tls-certs\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.812791 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-config-data\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.812862 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-logs\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.812939 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ntjr\" (UniqueName: \"kubernetes.io/projected/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-kube-api-access-8ntjr\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.812979 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.813020 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.813057 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-public-tls-certs\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.813909 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-logs\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.820397 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.820422 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-public-tls-certs\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.820563 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-config-data\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.820721 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.828715 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ntjr\" (UniqueName: \"kubernetes.io/projected/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-kube-api-access-8ntjr\") pod \"nova-api-0\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " pod="openstack/nova-api-0" Feb 25 13:41:09 crc kubenswrapper[4815]: I0225 13:41:09.947490 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 13:41:10 crc kubenswrapper[4815]: I0225 13:41:10.454267 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 13:41:10 crc kubenswrapper[4815]: W0225 13:41:10.458879 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9e0be7f_849b_4873_83fd_c0a440f5e9e3.slice/crio-0d8fec7030debdcca3791e3ce241d2cd1857930bf5ddc04b1843b2dfa1746c10 WatchSource:0}: Error finding container 0d8fec7030debdcca3791e3ce241d2cd1857930bf5ddc04b1843b2dfa1746c10: Status 404 returned error can't find the container with id 0d8fec7030debdcca3791e3ce241d2cd1857930bf5ddc04b1843b2dfa1746c10 Feb 25 13:41:10 crc kubenswrapper[4815]: I0225 13:41:10.554410 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c9e0be7f-849b-4873-83fd-c0a440f5e9e3","Type":"ContainerStarted","Data":"0d8fec7030debdcca3791e3ce241d2cd1857930bf5ddc04b1843b2dfa1746c10"} Feb 25 13:41:10 crc kubenswrapper[4815]: I0225 13:41:10.558825 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98c6009e-1597-4edc-a7d9-3987e946bd57","Type":"ContainerStarted","Data":"c062c417fbacd2241a880fc5b67650db4b431fb8983aaea869f1ce272090d89f"} Feb 25 13:41:10 crc kubenswrapper[4815]: I0225 13:41:10.558896 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98c6009e-1597-4edc-a7d9-3987e946bd57","Type":"ContainerStarted","Data":"4db32f3b712d39bf1531ae784b45de14998cbce8733a3116acee79adfd19d52c"} Feb 25 13:41:10 crc kubenswrapper[4815]: I0225 13:41:10.948958 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a22e2aa-a70a-4b98-add7-59a629664422" path="/var/lib/kubelet/pods/0a22e2aa-a70a-4b98-add7-59a629664422/volumes" Feb 25 13:41:11 crc kubenswrapper[4815]: I0225 13:41:11.084181 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:11 crc kubenswrapper[4815]: I0225 13:41:11.132715 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:11 crc kubenswrapper[4815]: I0225 13:41:11.570042 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c9e0be7f-849b-4873-83fd-c0a440f5e9e3","Type":"ContainerStarted","Data":"ee4e1ba080fcbf1998c4e3fedb9a52e9e11a4499f42a5d3f64b92e2708f50b06"} Feb 25 13:41:11 crc kubenswrapper[4815]: I0225 13:41:11.570092 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c9e0be7f-849b-4873-83fd-c0a440f5e9e3","Type":"ContainerStarted","Data":"e4e30b97e28bf2911af1355865e92193110986b66ffd8495e16b31de87a0677e"} Feb 25 13:41:11 crc kubenswrapper[4815]: I0225 13:41:11.571632 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98c6009e-1597-4edc-a7d9-3987e946bd57","Type":"ContainerStarted","Data":"d65c83d812b0faa5f0dca6a9686ecb23538d1fff393bd5d487e48ab6ac979aac"} Feb 25 13:41:11 crc kubenswrapper[4815]: I0225 13:41:11.588319 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.58830407 podStartE2EDuration="2.58830407s" podCreationTimestamp="2026-02-25 13:41:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:41:11.584222607 +0000 UTC m=+1229.385320661" watchObservedRunningTime="2026-02-25 13:41:11.58830407 +0000 UTC m=+1229.389402114" Feb 25 13:41:11 crc kubenswrapper[4815]: I0225 13:41:11.605196 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 25 13:41:11 crc kubenswrapper[4815]: I0225 13:41:11.850999 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-8bpf7"] Feb 25 13:41:11 crc kubenswrapper[4815]: I0225 13:41:11.853011 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8bpf7" Feb 25 13:41:11 crc kubenswrapper[4815]: I0225 13:41:11.856450 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 25 13:41:11 crc kubenswrapper[4815]: I0225 13:41:11.857382 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 25 13:41:11 crc kubenswrapper[4815]: I0225 13:41:11.906210 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-8bpf7"] Feb 25 13:41:11 crc kubenswrapper[4815]: I0225 13:41:11.977878 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-scripts\") pod \"nova-cell1-cell-mapping-8bpf7\" (UID: \"a752cbdd-832b-4635-84af-03cecc4af21e\") " pod="openstack/nova-cell1-cell-mapping-8bpf7" Feb 25 13:41:11 crc kubenswrapper[4815]: I0225 13:41:11.978026 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8bpf7\" (UID: \"a752cbdd-832b-4635-84af-03cecc4af21e\") " pod="openstack/nova-cell1-cell-mapping-8bpf7" Feb 25 13:41:11 crc kubenswrapper[4815]: I0225 13:41:11.978075 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-config-data\") pod \"nova-cell1-cell-mapping-8bpf7\" (UID: \"a752cbdd-832b-4635-84af-03cecc4af21e\") " pod="openstack/nova-cell1-cell-mapping-8bpf7" Feb 25 13:41:11 crc kubenswrapper[4815]: I0225 13:41:11.978097 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k8ft\" (UniqueName: \"kubernetes.io/projected/a752cbdd-832b-4635-84af-03cecc4af21e-kube-api-access-2k8ft\") pod \"nova-cell1-cell-mapping-8bpf7\" (UID: \"a752cbdd-832b-4635-84af-03cecc4af21e\") " pod="openstack/nova-cell1-cell-mapping-8bpf7" Feb 25 13:41:12 crc kubenswrapper[4815]: I0225 13:41:12.081097 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8bpf7\" (UID: \"a752cbdd-832b-4635-84af-03cecc4af21e\") " pod="openstack/nova-cell1-cell-mapping-8bpf7" Feb 25 13:41:12 crc kubenswrapper[4815]: I0225 13:41:12.081470 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-config-data\") pod \"nova-cell1-cell-mapping-8bpf7\" (UID: \"a752cbdd-832b-4635-84af-03cecc4af21e\") " pod="openstack/nova-cell1-cell-mapping-8bpf7" Feb 25 13:41:12 crc kubenswrapper[4815]: I0225 13:41:12.081529 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k8ft\" (UniqueName: \"kubernetes.io/projected/a752cbdd-832b-4635-84af-03cecc4af21e-kube-api-access-2k8ft\") pod \"nova-cell1-cell-mapping-8bpf7\" (UID: \"a752cbdd-832b-4635-84af-03cecc4af21e\") " pod="openstack/nova-cell1-cell-mapping-8bpf7" Feb 25 13:41:12 crc kubenswrapper[4815]: I0225 13:41:12.081615 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-scripts\") pod \"nova-cell1-cell-mapping-8bpf7\" (UID: \"a752cbdd-832b-4635-84af-03cecc4af21e\") " pod="openstack/nova-cell1-cell-mapping-8bpf7" Feb 25 13:41:12 crc kubenswrapper[4815]: I0225 13:41:12.089007 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8bpf7\" (UID: \"a752cbdd-832b-4635-84af-03cecc4af21e\") " pod="openstack/nova-cell1-cell-mapping-8bpf7" Feb 25 13:41:12 crc kubenswrapper[4815]: I0225 13:41:12.089241 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-config-data\") pod \"nova-cell1-cell-mapping-8bpf7\" (UID: \"a752cbdd-832b-4635-84af-03cecc4af21e\") " pod="openstack/nova-cell1-cell-mapping-8bpf7" Feb 25 13:41:12 crc kubenswrapper[4815]: I0225 13:41:12.093392 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-scripts\") pod \"nova-cell1-cell-mapping-8bpf7\" (UID: \"a752cbdd-832b-4635-84af-03cecc4af21e\") " pod="openstack/nova-cell1-cell-mapping-8bpf7" Feb 25 13:41:12 crc kubenswrapper[4815]: I0225 13:41:12.099731 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k8ft\" (UniqueName: \"kubernetes.io/projected/a752cbdd-832b-4635-84af-03cecc4af21e-kube-api-access-2k8ft\") pod \"nova-cell1-cell-mapping-8bpf7\" (UID: \"a752cbdd-832b-4635-84af-03cecc4af21e\") " pod="openstack/nova-cell1-cell-mapping-8bpf7" Feb 25 13:41:12 crc kubenswrapper[4815]: I0225 13:41:12.189142 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8bpf7" Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:12.583569 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98c6009e-1597-4edc-a7d9-3987e946bd57","Type":"ContainerStarted","Data":"0db2d9c4ca5865263fbb27ecb11e95ee6d1a534e229ab5b4823441ec2b15235a"} Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:12.637999 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-8bpf7"] Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:13.015681 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:13.084026 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-r49tp"] Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:13.084308 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-865f5d856f-r49tp" podUID="437f9211-e28f-4da9-993b-9a4dd71796f6" containerName="dnsmasq-dns" containerID="cri-o://9c8e4523cc93cae849c9317f02d216c204cd816fb738b3b6ead4dd808da01f5e" gracePeriod=10 Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:13.633840 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8bpf7" event={"ID":"a752cbdd-832b-4635-84af-03cecc4af21e","Type":"ContainerStarted","Data":"e97e5bb19afbf571975e89fe8a65df1145d57e6c6b03e8207d710fb992a8aa33"} Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:13.634191 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8bpf7" event={"ID":"a752cbdd-832b-4635-84af-03cecc4af21e","Type":"ContainerStarted","Data":"161fc7d60995188435ebac1c02c6ee5200826f2e29649359917411be90edaacb"} Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:13.664779 4815 generic.go:334] "Generic (PLEG): container finished" podID="437f9211-e28f-4da9-993b-9a4dd71796f6" containerID="9c8e4523cc93cae849c9317f02d216c204cd816fb738b3b6ead4dd808da01f5e" exitCode=0 Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:13.664824 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-r49tp" event={"ID":"437f9211-e28f-4da9-993b-9a4dd71796f6","Type":"ContainerDied","Data":"9c8e4523cc93cae849c9317f02d216c204cd816fb738b3b6ead4dd808da01f5e"} Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:13.664851 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-r49tp" event={"ID":"437f9211-e28f-4da9-993b-9a4dd71796f6","Type":"ContainerDied","Data":"851e7f720f745965982651882048f37651e84c1620e69de60c30f8b474eabfcb"} Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:13.664863 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="851e7f720f745965982651882048f37651e84c1620e69de60c30f8b474eabfcb" Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:13.666019 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:13.713801 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-config\") pod \"437f9211-e28f-4da9-993b-9a4dd71796f6\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:13.713851 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-ovsdbserver-nb\") pod \"437f9211-e28f-4da9-993b-9a4dd71796f6\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:13.713872 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9lpp\" (UniqueName: \"kubernetes.io/projected/437f9211-e28f-4da9-993b-9a4dd71796f6-kube-api-access-n9lpp\") pod \"437f9211-e28f-4da9-993b-9a4dd71796f6\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:13.713896 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-dns-swift-storage-0\") pod \"437f9211-e28f-4da9-993b-9a4dd71796f6\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:13.713954 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-ovsdbserver-sb\") pod \"437f9211-e28f-4da9-993b-9a4dd71796f6\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " Feb 25 13:41:13 crc kubenswrapper[4815]: I0225 13:41:13.714021 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-dns-svc\") pod \"437f9211-e28f-4da9-993b-9a4dd71796f6\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " Feb 25 13:41:14 crc kubenswrapper[4815]: I0225 13:41:14.908636 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/437f9211-e28f-4da9-993b-9a4dd71796f6-kube-api-access-n9lpp" (OuterVolumeSpecName: "kube-api-access-n9lpp") pod "437f9211-e28f-4da9-993b-9a4dd71796f6" (UID: "437f9211-e28f-4da9-993b-9a4dd71796f6"). InnerVolumeSpecName "kube-api-access-n9lpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:41:14 crc kubenswrapper[4815]: I0225 13:41:14.941129 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "437f9211-e28f-4da9-993b-9a4dd71796f6" (UID: "437f9211-e28f-4da9-993b-9a4dd71796f6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:41:14 crc kubenswrapper[4815]: I0225 13:41:14.946342 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-r49tp" Feb 25 13:41:14 crc kubenswrapper[4815]: I0225 13:41:14.952906 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-dns-swift-storage-0\") pod \"437f9211-e28f-4da9-993b-9a4dd71796f6\" (UID: \"437f9211-e28f-4da9-993b-9a4dd71796f6\") " Feb 25 13:41:14 crc kubenswrapper[4815]: I0225 13:41:14.963871 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "437f9211-e28f-4da9-993b-9a4dd71796f6" (UID: "437f9211-e28f-4da9-993b-9a4dd71796f6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:41:14 crc kubenswrapper[4815]: W0225 13:41:14.970665 4815 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/437f9211-e28f-4da9-993b-9a4dd71796f6/volumes/kubernetes.io~configmap/dns-swift-storage-0 Feb 25 13:41:14 crc kubenswrapper[4815]: I0225 13:41:14.970699 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "437f9211-e28f-4da9-993b-9a4dd71796f6" (UID: "437f9211-e28f-4da9-993b-9a4dd71796f6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:41:15 crc kubenswrapper[4815]: I0225 13:41:14.985654 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-8bpf7" podStartSLOduration=3.985633785 podStartE2EDuration="3.985633785s" podCreationTimestamp="2026-02-25 13:41:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:41:14.930357315 +0000 UTC m=+1232.731455369" watchObservedRunningTime="2026-02-25 13:41:14.985633785 +0000 UTC m=+1232.786731839" Feb 25 13:41:15 crc kubenswrapper[4815]: I0225 13:41:15.010653 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-config" (OuterVolumeSpecName: "config") pod "437f9211-e28f-4da9-993b-9a4dd71796f6" (UID: "437f9211-e28f-4da9-993b-9a4dd71796f6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:41:15 crc kubenswrapper[4815]: I0225 13:41:15.018632 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9lpp\" (UniqueName: \"kubernetes.io/projected/437f9211-e28f-4da9-993b-9a4dd71796f6-kube-api-access-n9lpp\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:15 crc kubenswrapper[4815]: I0225 13:41:15.018670 4815 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:15 crc kubenswrapper[4815]: I0225 13:41:15.054453 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "437f9211-e28f-4da9-993b-9a4dd71796f6" (UID: "437f9211-e28f-4da9-993b-9a4dd71796f6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:41:15 crc kubenswrapper[4815]: I0225 13:41:15.106862 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "437f9211-e28f-4da9-993b-9a4dd71796f6" (UID: "437f9211-e28f-4da9-993b-9a4dd71796f6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:41:15 crc kubenswrapper[4815]: I0225 13:41:15.120404 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:15 crc kubenswrapper[4815]: I0225 13:41:15.120426 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:15 crc kubenswrapper[4815]: I0225 13:41:15.120436 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:15 crc kubenswrapper[4815]: I0225 13:41:15.120444 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/437f9211-e28f-4da9-993b-9a4dd71796f6-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:15 crc kubenswrapper[4815]: I0225 13:41:15.295999 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-r49tp"] Feb 25 13:41:15 crc kubenswrapper[4815]: I0225 13:41:15.304313 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-r49tp"] Feb 25 13:41:15 crc kubenswrapper[4815]: I0225 13:41:15.958067 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98c6009e-1597-4edc-a7d9-3987e946bd57","Type":"ContainerStarted","Data":"2d18c97d0684d868ce92668cdd759073fd978ee73abbe83f1f8c20b019b951b0"} Feb 25 13:41:15 crc kubenswrapper[4815]: I0225 13:41:15.959979 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 13:41:15 crc kubenswrapper[4815]: I0225 13:41:15.996318 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.692236173 podStartE2EDuration="7.996294321s" podCreationTimestamp="2026-02-25 13:41:08 +0000 UTC" firstStartedPulling="2026-02-25 13:41:09.653749997 +0000 UTC m=+1227.454848051" lastFinishedPulling="2026-02-25 13:41:14.957808145 +0000 UTC m=+1232.758906199" observedRunningTime="2026-02-25 13:41:15.992148934 +0000 UTC m=+1233.793247028" watchObservedRunningTime="2026-02-25 13:41:15.996294321 +0000 UTC m=+1233.797392415" Feb 25 13:41:16 crc kubenswrapper[4815]: I0225 13:41:16.953792 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="437f9211-e28f-4da9-993b-9a4dd71796f6" path="/var/lib/kubelet/pods/437f9211-e28f-4da9-993b-9a4dd71796f6/volumes" Feb 25 13:41:17 crc kubenswrapper[4815]: I0225 13:41:17.995849 4815 generic.go:334] "Generic (PLEG): container finished" podID="a752cbdd-832b-4635-84af-03cecc4af21e" containerID="e97e5bb19afbf571975e89fe8a65df1145d57e6c6b03e8207d710fb992a8aa33" exitCode=0 Feb 25 13:41:17 crc kubenswrapper[4815]: I0225 13:41:17.995968 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8bpf7" event={"ID":"a752cbdd-832b-4635-84af-03cecc4af21e","Type":"ContainerDied","Data":"e97e5bb19afbf571975e89fe8a65df1145d57e6c6b03e8207d710fb992a8aa33"} Feb 25 13:41:19 crc kubenswrapper[4815]: I0225 13:41:19.418078 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8bpf7" Feb 25 13:41:19 crc kubenswrapper[4815]: I0225 13:41:19.605808 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-config-data\") pod \"a752cbdd-832b-4635-84af-03cecc4af21e\" (UID: \"a752cbdd-832b-4635-84af-03cecc4af21e\") " Feb 25 13:41:19 crc kubenswrapper[4815]: I0225 13:41:19.606231 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-scripts\") pod \"a752cbdd-832b-4635-84af-03cecc4af21e\" (UID: \"a752cbdd-832b-4635-84af-03cecc4af21e\") " Feb 25 13:41:19 crc kubenswrapper[4815]: I0225 13:41:19.606337 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-combined-ca-bundle\") pod \"a752cbdd-832b-4635-84af-03cecc4af21e\" (UID: \"a752cbdd-832b-4635-84af-03cecc4af21e\") " Feb 25 13:41:19 crc kubenswrapper[4815]: I0225 13:41:19.606762 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k8ft\" (UniqueName: \"kubernetes.io/projected/a752cbdd-832b-4635-84af-03cecc4af21e-kube-api-access-2k8ft\") pod \"a752cbdd-832b-4635-84af-03cecc4af21e\" (UID: \"a752cbdd-832b-4635-84af-03cecc4af21e\") " Feb 25 13:41:19 crc kubenswrapper[4815]: I0225 13:41:19.614178 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-scripts" (OuterVolumeSpecName: "scripts") pod "a752cbdd-832b-4635-84af-03cecc4af21e" (UID: "a752cbdd-832b-4635-84af-03cecc4af21e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:19 crc kubenswrapper[4815]: I0225 13:41:19.614393 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a752cbdd-832b-4635-84af-03cecc4af21e-kube-api-access-2k8ft" (OuterVolumeSpecName: "kube-api-access-2k8ft") pod "a752cbdd-832b-4635-84af-03cecc4af21e" (UID: "a752cbdd-832b-4635-84af-03cecc4af21e"). InnerVolumeSpecName "kube-api-access-2k8ft". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:41:19 crc kubenswrapper[4815]: I0225 13:41:19.644764 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a752cbdd-832b-4635-84af-03cecc4af21e" (UID: "a752cbdd-832b-4635-84af-03cecc4af21e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:19 crc kubenswrapper[4815]: I0225 13:41:19.663730 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-config-data" (OuterVolumeSpecName: "config-data") pod "a752cbdd-832b-4635-84af-03cecc4af21e" (UID: "a752cbdd-832b-4635-84af-03cecc4af21e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:19 crc kubenswrapper[4815]: I0225 13:41:19.709108 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k8ft\" (UniqueName: \"kubernetes.io/projected/a752cbdd-832b-4635-84af-03cecc4af21e-kube-api-access-2k8ft\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:19 crc kubenswrapper[4815]: I0225 13:41:19.709147 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:19 crc kubenswrapper[4815]: I0225 13:41:19.709164 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:19 crc kubenswrapper[4815]: I0225 13:41:19.709178 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a752cbdd-832b-4635-84af-03cecc4af21e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:19 crc kubenswrapper[4815]: I0225 13:41:19.948594 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 13:41:19 crc kubenswrapper[4815]: I0225 13:41:19.948682 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 13:41:20 crc kubenswrapper[4815]: I0225 13:41:20.022393 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8bpf7" event={"ID":"a752cbdd-832b-4635-84af-03cecc4af21e","Type":"ContainerDied","Data":"161fc7d60995188435ebac1c02c6ee5200826f2e29649359917411be90edaacb"} Feb 25 13:41:20 crc kubenswrapper[4815]: I0225 13:41:20.022449 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="161fc7d60995188435ebac1c02c6ee5200826f2e29649359917411be90edaacb" Feb 25 13:41:20 crc kubenswrapper[4815]: I0225 13:41:20.022620 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8bpf7" Feb 25 13:41:20 crc kubenswrapper[4815]: I0225 13:41:20.234792 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 13:41:20 crc kubenswrapper[4815]: I0225 13:41:20.235081 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c9e0be7f-849b-4873-83fd-c0a440f5e9e3" containerName="nova-api-log" containerID="cri-o://e4e30b97e28bf2911af1355865e92193110986b66ffd8495e16b31de87a0677e" gracePeriod=30 Feb 25 13:41:20 crc kubenswrapper[4815]: I0225 13:41:20.235210 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c9e0be7f-849b-4873-83fd-c0a440f5e9e3" containerName="nova-api-api" containerID="cri-o://ee4e1ba080fcbf1998c4e3fedb9a52e9e11a4499f42a5d3f64b92e2708f50b06" gracePeriod=30 Feb 25 13:41:20 crc kubenswrapper[4815]: I0225 13:41:20.240720 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c9e0be7f-849b-4873-83fd-c0a440f5e9e3" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": EOF" Feb 25 13:41:20 crc kubenswrapper[4815]: I0225 13:41:20.240936 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c9e0be7f-849b-4873-83fd-c0a440f5e9e3" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": EOF" Feb 25 13:41:20 crc kubenswrapper[4815]: I0225 13:41:20.289783 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 13:41:20 crc kubenswrapper[4815]: I0225 13:41:20.290138 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="86841bf5-3d74-480b-b7b4-ffb6534fb441" containerName="nova-scheduler-scheduler" containerID="cri-o://f128963912d5f34faf0049947b1af6fc5870933479b5c5de0d1791f6da26f212" gracePeriod=30 Feb 25 13:41:20 crc kubenswrapper[4815]: I0225 13:41:20.304013 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:41:20 crc kubenswrapper[4815]: I0225 13:41:20.304285 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="302dd34a-26ad-4c26-aa17-fe902863fd4b" containerName="nova-metadata-log" containerID="cri-o://65d029120089702d344365518d9dc68dc834a9c72765245c25c72ca180c5dd91" gracePeriod=30 Feb 25 13:41:20 crc kubenswrapper[4815]: I0225 13:41:20.304451 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="302dd34a-26ad-4c26-aa17-fe902863fd4b" containerName="nova-metadata-metadata" containerID="cri-o://3d51ebbf50ce3f86ae89020d264c8ca488baa04f458010b1dd140ccca80f28a1" gracePeriod=30 Feb 25 13:41:21 crc kubenswrapper[4815]: I0225 13:41:21.035487 4815 generic.go:334] "Generic (PLEG): container finished" podID="c9e0be7f-849b-4873-83fd-c0a440f5e9e3" containerID="e4e30b97e28bf2911af1355865e92193110986b66ffd8495e16b31de87a0677e" exitCode=143 Feb 25 13:41:21 crc kubenswrapper[4815]: I0225 13:41:21.035549 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c9e0be7f-849b-4873-83fd-c0a440f5e9e3","Type":"ContainerDied","Data":"e4e30b97e28bf2911af1355865e92193110986b66ffd8495e16b31de87a0677e"} Feb 25 13:41:21 crc kubenswrapper[4815]: I0225 13:41:21.038173 4815 generic.go:334] "Generic (PLEG): container finished" podID="302dd34a-26ad-4c26-aa17-fe902863fd4b" containerID="65d029120089702d344365518d9dc68dc834a9c72765245c25c72ca180c5dd91" exitCode=143 Feb 25 13:41:21 crc kubenswrapper[4815]: I0225 13:41:21.038221 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"302dd34a-26ad-4c26-aa17-fe902863fd4b","Type":"ContainerDied","Data":"65d029120089702d344365518d9dc68dc834a9c72765245c25c72ca180c5dd91"} Feb 25 13:41:21 crc kubenswrapper[4815]: E0225 13:41:21.899591 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f128963912d5f34faf0049947b1af6fc5870933479b5c5de0d1791f6da26f212" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 13:41:21 crc kubenswrapper[4815]: E0225 13:41:21.901536 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f128963912d5f34faf0049947b1af6fc5870933479b5c5de0d1791f6da26f212" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 13:41:21 crc kubenswrapper[4815]: E0225 13:41:21.903899 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f128963912d5f34faf0049947b1af6fc5870933479b5c5de0d1791f6da26f212" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 13:41:21 crc kubenswrapper[4815]: E0225 13:41:21.903946 4815 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="86841bf5-3d74-480b-b7b4-ffb6534fb441" containerName="nova-scheduler-scheduler" Feb 25 13:41:23 crc kubenswrapper[4815]: I0225 13:41:23.451329 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="302dd34a-26ad-4c26-aa17-fe902863fd4b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": read tcp 10.217.0.2:35086->10.217.0.206:8775: read: connection reset by peer" Feb 25 13:41:23 crc kubenswrapper[4815]: I0225 13:41:23.451424 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="302dd34a-26ad-4c26-aa17-fe902863fd4b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": read tcp 10.217.0.2:35084->10.217.0.206:8775: read: connection reset by peer" Feb 25 13:41:23 crc kubenswrapper[4815]: I0225 13:41:23.901357 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.018316 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/302dd34a-26ad-4c26-aa17-fe902863fd4b-logs\") pod \"302dd34a-26ad-4c26-aa17-fe902863fd4b\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.018409 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-nova-metadata-tls-certs\") pod \"302dd34a-26ad-4c26-aa17-fe902863fd4b\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.018614 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-combined-ca-bundle\") pod \"302dd34a-26ad-4c26-aa17-fe902863fd4b\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.018748 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-config-data\") pod \"302dd34a-26ad-4c26-aa17-fe902863fd4b\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.018827 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cf5qf\" (UniqueName: \"kubernetes.io/projected/302dd34a-26ad-4c26-aa17-fe902863fd4b-kube-api-access-cf5qf\") pod \"302dd34a-26ad-4c26-aa17-fe902863fd4b\" (UID: \"302dd34a-26ad-4c26-aa17-fe902863fd4b\") " Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.020150 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/302dd34a-26ad-4c26-aa17-fe902863fd4b-logs" (OuterVolumeSpecName: "logs") pod "302dd34a-26ad-4c26-aa17-fe902863fd4b" (UID: "302dd34a-26ad-4c26-aa17-fe902863fd4b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.061105 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/302dd34a-26ad-4c26-aa17-fe902863fd4b-kube-api-access-cf5qf" (OuterVolumeSpecName: "kube-api-access-cf5qf") pod "302dd34a-26ad-4c26-aa17-fe902863fd4b" (UID: "302dd34a-26ad-4c26-aa17-fe902863fd4b"). InnerVolumeSpecName "kube-api-access-cf5qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.066598 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-config-data" (OuterVolumeSpecName: "config-data") pod "302dd34a-26ad-4c26-aa17-fe902863fd4b" (UID: "302dd34a-26ad-4c26-aa17-fe902863fd4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.069054 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "302dd34a-26ad-4c26-aa17-fe902863fd4b" (UID: "302dd34a-26ad-4c26-aa17-fe902863fd4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.076607 4815 generic.go:334] "Generic (PLEG): container finished" podID="302dd34a-26ad-4c26-aa17-fe902863fd4b" containerID="3d51ebbf50ce3f86ae89020d264c8ca488baa04f458010b1dd140ccca80f28a1" exitCode=0 Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.076658 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"302dd34a-26ad-4c26-aa17-fe902863fd4b","Type":"ContainerDied","Data":"3d51ebbf50ce3f86ae89020d264c8ca488baa04f458010b1dd140ccca80f28a1"} Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.076689 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"302dd34a-26ad-4c26-aa17-fe902863fd4b","Type":"ContainerDied","Data":"11e35352b80a044ae4b0b6fadbb7f3c5b6d9e9e7dc5a08e2cb31a1f61fdac0a8"} Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.076710 4815 scope.go:117] "RemoveContainer" containerID="3d51ebbf50ce3f86ae89020d264c8ca488baa04f458010b1dd140ccca80f28a1" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.076856 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.091415 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "302dd34a-26ad-4c26-aa17-fe902863fd4b" (UID: "302dd34a-26ad-4c26-aa17-fe902863fd4b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.121726 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.121766 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.121782 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cf5qf\" (UniqueName: \"kubernetes.io/projected/302dd34a-26ad-4c26-aa17-fe902863fd4b-kube-api-access-cf5qf\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.121797 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/302dd34a-26ad-4c26-aa17-fe902863fd4b-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.121811 4815 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/302dd34a-26ad-4c26-aa17-fe902863fd4b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.141292 4815 scope.go:117] "RemoveContainer" containerID="65d029120089702d344365518d9dc68dc834a9c72765245c25c72ca180c5dd91" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.159684 4815 scope.go:117] "RemoveContainer" containerID="3d51ebbf50ce3f86ae89020d264c8ca488baa04f458010b1dd140ccca80f28a1" Feb 25 13:41:24 crc kubenswrapper[4815]: E0225 13:41:24.160114 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d51ebbf50ce3f86ae89020d264c8ca488baa04f458010b1dd140ccca80f28a1\": container with ID starting with 3d51ebbf50ce3f86ae89020d264c8ca488baa04f458010b1dd140ccca80f28a1 not found: ID does not exist" containerID="3d51ebbf50ce3f86ae89020d264c8ca488baa04f458010b1dd140ccca80f28a1" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.160216 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d51ebbf50ce3f86ae89020d264c8ca488baa04f458010b1dd140ccca80f28a1"} err="failed to get container status \"3d51ebbf50ce3f86ae89020d264c8ca488baa04f458010b1dd140ccca80f28a1\": rpc error: code = NotFound desc = could not find container \"3d51ebbf50ce3f86ae89020d264c8ca488baa04f458010b1dd140ccca80f28a1\": container with ID starting with 3d51ebbf50ce3f86ae89020d264c8ca488baa04f458010b1dd140ccca80f28a1 not found: ID does not exist" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.160310 4815 scope.go:117] "RemoveContainer" containerID="65d029120089702d344365518d9dc68dc834a9c72765245c25c72ca180c5dd91" Feb 25 13:41:24 crc kubenswrapper[4815]: E0225 13:41:24.160656 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65d029120089702d344365518d9dc68dc834a9c72765245c25c72ca180c5dd91\": container with ID starting with 65d029120089702d344365518d9dc68dc834a9c72765245c25c72ca180c5dd91 not found: ID does not exist" containerID="65d029120089702d344365518d9dc68dc834a9c72765245c25c72ca180c5dd91" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.160694 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65d029120089702d344365518d9dc68dc834a9c72765245c25c72ca180c5dd91"} err="failed to get container status \"65d029120089702d344365518d9dc68dc834a9c72765245c25c72ca180c5dd91\": rpc error: code = NotFound desc = could not find container \"65d029120089702d344365518d9dc68dc834a9c72765245c25c72ca180c5dd91\": container with ID starting with 65d029120089702d344365518d9dc68dc834a9c72765245c25c72ca180c5dd91 not found: ID does not exist" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.413003 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.426742 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.438728 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:41:24 crc kubenswrapper[4815]: E0225 13:41:24.439205 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="437f9211-e28f-4da9-993b-9a4dd71796f6" containerName="init" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.439229 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="437f9211-e28f-4da9-993b-9a4dd71796f6" containerName="init" Feb 25 13:41:24 crc kubenswrapper[4815]: E0225 13:41:24.439253 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a752cbdd-832b-4635-84af-03cecc4af21e" containerName="nova-manage" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.439262 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a752cbdd-832b-4635-84af-03cecc4af21e" containerName="nova-manage" Feb 25 13:41:24 crc kubenswrapper[4815]: E0225 13:41:24.439276 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="302dd34a-26ad-4c26-aa17-fe902863fd4b" containerName="nova-metadata-metadata" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.439285 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="302dd34a-26ad-4c26-aa17-fe902863fd4b" containerName="nova-metadata-metadata" Feb 25 13:41:24 crc kubenswrapper[4815]: E0225 13:41:24.439302 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="437f9211-e28f-4da9-993b-9a4dd71796f6" containerName="dnsmasq-dns" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.439310 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="437f9211-e28f-4da9-993b-9a4dd71796f6" containerName="dnsmasq-dns" Feb 25 13:41:24 crc kubenswrapper[4815]: E0225 13:41:24.439351 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="302dd34a-26ad-4c26-aa17-fe902863fd4b" containerName="nova-metadata-log" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.439361 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="302dd34a-26ad-4c26-aa17-fe902863fd4b" containerName="nova-metadata-log" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.439612 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="a752cbdd-832b-4635-84af-03cecc4af21e" containerName="nova-manage" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.439631 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="302dd34a-26ad-4c26-aa17-fe902863fd4b" containerName="nova-metadata-metadata" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.439669 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="302dd34a-26ad-4c26-aa17-fe902863fd4b" containerName="nova-metadata-log" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.439683 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="437f9211-e28f-4da9-993b-9a4dd71796f6" containerName="dnsmasq-dns" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.440887 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.444122 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.447763 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.454419 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.529583 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20eee08-f0d7-49bb-a195-202a600195f9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d20eee08-f0d7-49bb-a195-202a600195f9\") " pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.529630 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hszls\" (UniqueName: \"kubernetes.io/projected/d20eee08-f0d7-49bb-a195-202a600195f9-kube-api-access-hszls\") pod \"nova-metadata-0\" (UID: \"d20eee08-f0d7-49bb-a195-202a600195f9\") " pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.529655 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d20eee08-f0d7-49bb-a195-202a600195f9-logs\") pod \"nova-metadata-0\" (UID: \"d20eee08-f0d7-49bb-a195-202a600195f9\") " pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.529673 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d20eee08-f0d7-49bb-a195-202a600195f9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d20eee08-f0d7-49bb-a195-202a600195f9\") " pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.529771 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20eee08-f0d7-49bb-a195-202a600195f9-config-data\") pod \"nova-metadata-0\" (UID: \"d20eee08-f0d7-49bb-a195-202a600195f9\") " pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.631985 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20eee08-f0d7-49bb-a195-202a600195f9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d20eee08-f0d7-49bb-a195-202a600195f9\") " pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.632067 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hszls\" (UniqueName: \"kubernetes.io/projected/d20eee08-f0d7-49bb-a195-202a600195f9-kube-api-access-hszls\") pod \"nova-metadata-0\" (UID: \"d20eee08-f0d7-49bb-a195-202a600195f9\") " pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.632102 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d20eee08-f0d7-49bb-a195-202a600195f9-logs\") pod \"nova-metadata-0\" (UID: \"d20eee08-f0d7-49bb-a195-202a600195f9\") " pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.632131 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d20eee08-f0d7-49bb-a195-202a600195f9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d20eee08-f0d7-49bb-a195-202a600195f9\") " pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.632273 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20eee08-f0d7-49bb-a195-202a600195f9-config-data\") pod \"nova-metadata-0\" (UID: \"d20eee08-f0d7-49bb-a195-202a600195f9\") " pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.632685 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d20eee08-f0d7-49bb-a195-202a600195f9-logs\") pod \"nova-metadata-0\" (UID: \"d20eee08-f0d7-49bb-a195-202a600195f9\") " pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.636779 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d20eee08-f0d7-49bb-a195-202a600195f9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d20eee08-f0d7-49bb-a195-202a600195f9\") " pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.637455 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d20eee08-f0d7-49bb-a195-202a600195f9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d20eee08-f0d7-49bb-a195-202a600195f9\") " pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.640198 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d20eee08-f0d7-49bb-a195-202a600195f9-config-data\") pod \"nova-metadata-0\" (UID: \"d20eee08-f0d7-49bb-a195-202a600195f9\") " pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.651680 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hszls\" (UniqueName: \"kubernetes.io/projected/d20eee08-f0d7-49bb-a195-202a600195f9-kube-api-access-hszls\") pod \"nova-metadata-0\" (UID: \"d20eee08-f0d7-49bb-a195-202a600195f9\") " pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.776783 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 13:41:24 crc kubenswrapper[4815]: I0225 13:41:24.952237 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="302dd34a-26ad-4c26-aa17-fe902863fd4b" path="/var/lib/kubelet/pods/302dd34a-26ad-4c26-aa17-fe902863fd4b/volumes" Feb 25 13:41:25 crc kubenswrapper[4815]: I0225 13:41:25.308826 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 13:41:25 crc kubenswrapper[4815]: W0225 13:41:25.350318 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd20eee08_f0d7_49bb_a195_202a600195f9.slice/crio-ad244fc1b048423ccd8f96b34056d5de1b1cafca9eabd454214f7fa1d015ea0f WatchSource:0}: Error finding container ad244fc1b048423ccd8f96b34056d5de1b1cafca9eabd454214f7fa1d015ea0f: Status 404 returned error can't find the container with id ad244fc1b048423ccd8f96b34056d5de1b1cafca9eabd454214f7fa1d015ea0f Feb 25 13:41:25 crc kubenswrapper[4815]: I0225 13:41:25.800229 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 13:41:25 crc kubenswrapper[4815]: I0225 13:41:25.856076 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmkmb\" (UniqueName: \"kubernetes.io/projected/86841bf5-3d74-480b-b7b4-ffb6534fb441-kube-api-access-bmkmb\") pod \"86841bf5-3d74-480b-b7b4-ffb6534fb441\" (UID: \"86841bf5-3d74-480b-b7b4-ffb6534fb441\") " Feb 25 13:41:25 crc kubenswrapper[4815]: I0225 13:41:25.856386 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86841bf5-3d74-480b-b7b4-ffb6534fb441-config-data\") pod \"86841bf5-3d74-480b-b7b4-ffb6534fb441\" (UID: \"86841bf5-3d74-480b-b7b4-ffb6534fb441\") " Feb 25 13:41:25 crc kubenswrapper[4815]: I0225 13:41:25.856541 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86841bf5-3d74-480b-b7b4-ffb6534fb441-combined-ca-bundle\") pod \"86841bf5-3d74-480b-b7b4-ffb6534fb441\" (UID: \"86841bf5-3d74-480b-b7b4-ffb6534fb441\") " Feb 25 13:41:25 crc kubenswrapper[4815]: I0225 13:41:25.879628 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86841bf5-3d74-480b-b7b4-ffb6534fb441-kube-api-access-bmkmb" (OuterVolumeSpecName: "kube-api-access-bmkmb") pod "86841bf5-3d74-480b-b7b4-ffb6534fb441" (UID: "86841bf5-3d74-480b-b7b4-ffb6534fb441"). InnerVolumeSpecName "kube-api-access-bmkmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:41:25 crc kubenswrapper[4815]: I0225 13:41:25.921310 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86841bf5-3d74-480b-b7b4-ffb6534fb441-config-data" (OuterVolumeSpecName: "config-data") pod "86841bf5-3d74-480b-b7b4-ffb6534fb441" (UID: "86841bf5-3d74-480b-b7b4-ffb6534fb441"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:25 crc kubenswrapper[4815]: I0225 13:41:25.938268 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86841bf5-3d74-480b-b7b4-ffb6534fb441-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86841bf5-3d74-480b-b7b4-ffb6534fb441" (UID: "86841bf5-3d74-480b-b7b4-ffb6534fb441"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:25 crc kubenswrapper[4815]: I0225 13:41:25.958841 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86841bf5-3d74-480b-b7b4-ffb6534fb441-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:25 crc kubenswrapper[4815]: I0225 13:41:25.959081 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmkmb\" (UniqueName: \"kubernetes.io/projected/86841bf5-3d74-480b-b7b4-ffb6534fb441-kube-api-access-bmkmb\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:25 crc kubenswrapper[4815]: I0225 13:41:25.959097 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86841bf5-3d74-480b-b7b4-ffb6534fb441-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:25 crc kubenswrapper[4815]: I0225 13:41:25.994327 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.060095 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-config-data\") pod \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.060165 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ntjr\" (UniqueName: \"kubernetes.io/projected/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-kube-api-access-8ntjr\") pod \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.060187 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-internal-tls-certs\") pod \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.060205 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-public-tls-certs\") pod \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.060292 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-combined-ca-bundle\") pod \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.060352 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-logs\") pod \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\" (UID: \"c9e0be7f-849b-4873-83fd-c0a440f5e9e3\") " Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.060920 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-logs" (OuterVolumeSpecName: "logs") pod "c9e0be7f-849b-4873-83fd-c0a440f5e9e3" (UID: "c9e0be7f-849b-4873-83fd-c0a440f5e9e3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.063985 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-kube-api-access-8ntjr" (OuterVolumeSpecName: "kube-api-access-8ntjr") pod "c9e0be7f-849b-4873-83fd-c0a440f5e9e3" (UID: "c9e0be7f-849b-4873-83fd-c0a440f5e9e3"). InnerVolumeSpecName "kube-api-access-8ntjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.088772 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c9e0be7f-849b-4873-83fd-c0a440f5e9e3" (UID: "c9e0be7f-849b-4873-83fd-c0a440f5e9e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.102630 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-config-data" (OuterVolumeSpecName: "config-data") pod "c9e0be7f-849b-4873-83fd-c0a440f5e9e3" (UID: "c9e0be7f-849b-4873-83fd-c0a440f5e9e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.104347 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d20eee08-f0d7-49bb-a195-202a600195f9","Type":"ContainerStarted","Data":"de37f204596f5c0abc4ecad005259980b262f9cc177ddf7372244d4789a3dc46"} Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.104389 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d20eee08-f0d7-49bb-a195-202a600195f9","Type":"ContainerStarted","Data":"1edc7da32e616cee0d1353ae11805dd533b74b919f91b700f294e7c593253886"} Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.104401 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d20eee08-f0d7-49bb-a195-202a600195f9","Type":"ContainerStarted","Data":"ad244fc1b048423ccd8f96b34056d5de1b1cafca9eabd454214f7fa1d015ea0f"} Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.110047 4815 generic.go:334] "Generic (PLEG): container finished" podID="c9e0be7f-849b-4873-83fd-c0a440f5e9e3" containerID="ee4e1ba080fcbf1998c4e3fedb9a52e9e11a4499f42a5d3f64b92e2708f50b06" exitCode=0 Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.110162 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c9e0be7f-849b-4873-83fd-c0a440f5e9e3","Type":"ContainerDied","Data":"ee4e1ba080fcbf1998c4e3fedb9a52e9e11a4499f42a5d3f64b92e2708f50b06"} Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.110194 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c9e0be7f-849b-4873-83fd-c0a440f5e9e3","Type":"ContainerDied","Data":"0d8fec7030debdcca3791e3ce241d2cd1857930bf5ddc04b1843b2dfa1746c10"} Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.110213 4815 scope.go:117] "RemoveContainer" containerID="ee4e1ba080fcbf1998c4e3fedb9a52e9e11a4499f42a5d3f64b92e2708f50b06" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.110400 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.112661 4815 generic.go:334] "Generic (PLEG): container finished" podID="86841bf5-3d74-480b-b7b4-ffb6534fb441" containerID="f128963912d5f34faf0049947b1af6fc5870933479b5c5de0d1791f6da26f212" exitCode=0 Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.112703 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"86841bf5-3d74-480b-b7b4-ffb6534fb441","Type":"ContainerDied","Data":"f128963912d5f34faf0049947b1af6fc5870933479b5c5de0d1791f6da26f212"} Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.112729 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"86841bf5-3d74-480b-b7b4-ffb6534fb441","Type":"ContainerDied","Data":"99e19d712ffce1dd44b94d8510b89b7b86676c84a2394e828ca790fecc44278b"} Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.112776 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.114157 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c9e0be7f-849b-4873-83fd-c0a440f5e9e3" (UID: "c9e0be7f-849b-4873-83fd-c0a440f5e9e3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.119367 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c9e0be7f-849b-4873-83fd-c0a440f5e9e3" (UID: "c9e0be7f-849b-4873-83fd-c0a440f5e9e3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.136829 4815 scope.go:117] "RemoveContainer" containerID="e4e30b97e28bf2911af1355865e92193110986b66ffd8495e16b31de87a0677e" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.145097 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.1450732869999998 podStartE2EDuration="2.145073287s" podCreationTimestamp="2026-02-25 13:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:41:26.138424014 +0000 UTC m=+1243.939522078" watchObservedRunningTime="2026-02-25 13:41:26.145073287 +0000 UTC m=+1243.946171341" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.159869 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.161445 4815 scope.go:117] "RemoveContainer" containerID="ee4e1ba080fcbf1998c4e3fedb9a52e9e11a4499f42a5d3f64b92e2708f50b06" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.163067 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.163090 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-logs\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.163102 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.163113 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ntjr\" (UniqueName: \"kubernetes.io/projected/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-kube-api-access-8ntjr\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.163123 4815 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.163131 4815 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9e0be7f-849b-4873-83fd-c0a440f5e9e3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 13:41:26 crc kubenswrapper[4815]: E0225 13:41:26.164595 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee4e1ba080fcbf1998c4e3fedb9a52e9e11a4499f42a5d3f64b92e2708f50b06\": container with ID starting with ee4e1ba080fcbf1998c4e3fedb9a52e9e11a4499f42a5d3f64b92e2708f50b06 not found: ID does not exist" containerID="ee4e1ba080fcbf1998c4e3fedb9a52e9e11a4499f42a5d3f64b92e2708f50b06" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.164632 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee4e1ba080fcbf1998c4e3fedb9a52e9e11a4499f42a5d3f64b92e2708f50b06"} err="failed to get container status \"ee4e1ba080fcbf1998c4e3fedb9a52e9e11a4499f42a5d3f64b92e2708f50b06\": rpc error: code = NotFound desc = could not find container \"ee4e1ba080fcbf1998c4e3fedb9a52e9e11a4499f42a5d3f64b92e2708f50b06\": container with ID starting with ee4e1ba080fcbf1998c4e3fedb9a52e9e11a4499f42a5d3f64b92e2708f50b06 not found: ID does not exist" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.164655 4815 scope.go:117] "RemoveContainer" containerID="e4e30b97e28bf2911af1355865e92193110986b66ffd8495e16b31de87a0677e" Feb 25 13:41:26 crc kubenswrapper[4815]: E0225 13:41:26.166868 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4e30b97e28bf2911af1355865e92193110986b66ffd8495e16b31de87a0677e\": container with ID starting with e4e30b97e28bf2911af1355865e92193110986b66ffd8495e16b31de87a0677e not found: ID does not exist" containerID="e4e30b97e28bf2911af1355865e92193110986b66ffd8495e16b31de87a0677e" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.166889 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4e30b97e28bf2911af1355865e92193110986b66ffd8495e16b31de87a0677e"} err="failed to get container status \"e4e30b97e28bf2911af1355865e92193110986b66ffd8495e16b31de87a0677e\": rpc error: code = NotFound desc = could not find container \"e4e30b97e28bf2911af1355865e92193110986b66ffd8495e16b31de87a0677e\": container with ID starting with e4e30b97e28bf2911af1355865e92193110986b66ffd8495e16b31de87a0677e not found: ID does not exist" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.166901 4815 scope.go:117] "RemoveContainer" containerID="f128963912d5f34faf0049947b1af6fc5870933479b5c5de0d1791f6da26f212" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.173471 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.187916 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 13:41:26 crc kubenswrapper[4815]: E0225 13:41:26.188296 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e0be7f-849b-4873-83fd-c0a440f5e9e3" containerName="nova-api-log" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.188308 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e0be7f-849b-4873-83fd-c0a440f5e9e3" containerName="nova-api-log" Feb 25 13:41:26 crc kubenswrapper[4815]: E0225 13:41:26.188324 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e0be7f-849b-4873-83fd-c0a440f5e9e3" containerName="nova-api-api" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.188330 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e0be7f-849b-4873-83fd-c0a440f5e9e3" containerName="nova-api-api" Feb 25 13:41:26 crc kubenswrapper[4815]: E0225 13:41:26.188337 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86841bf5-3d74-480b-b7b4-ffb6534fb441" containerName="nova-scheduler-scheduler" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.188343 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="86841bf5-3d74-480b-b7b4-ffb6534fb441" containerName="nova-scheduler-scheduler" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.188529 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="86841bf5-3d74-480b-b7b4-ffb6534fb441" containerName="nova-scheduler-scheduler" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.188544 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9e0be7f-849b-4873-83fd-c0a440f5e9e3" containerName="nova-api-api" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.188565 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9e0be7f-849b-4873-83fd-c0a440f5e9e3" containerName="nova-api-log" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.189126 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.192354 4815 scope.go:117] "RemoveContainer" containerID="f128963912d5f34faf0049947b1af6fc5870933479b5c5de0d1791f6da26f212" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.192560 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 25 13:41:26 crc kubenswrapper[4815]: E0225 13:41:26.193085 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f128963912d5f34faf0049947b1af6fc5870933479b5c5de0d1791f6da26f212\": container with ID starting with f128963912d5f34faf0049947b1af6fc5870933479b5c5de0d1791f6da26f212 not found: ID does not exist" containerID="f128963912d5f34faf0049947b1af6fc5870933479b5c5de0d1791f6da26f212" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.193119 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f128963912d5f34faf0049947b1af6fc5870933479b5c5de0d1791f6da26f212"} err="failed to get container status \"f128963912d5f34faf0049947b1af6fc5870933479b5c5de0d1791f6da26f212\": rpc error: code = NotFound desc = could not find container \"f128963912d5f34faf0049947b1af6fc5870933479b5c5de0d1791f6da26f212\": container with ID starting with f128963912d5f34faf0049947b1af6fc5870933479b5c5de0d1791f6da26f212 not found: ID does not exist" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.212798 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.265584 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v8j7\" (UniqueName: \"kubernetes.io/projected/441334b1-15af-4e17-8974-c5ab19338eee-kube-api-access-8v8j7\") pod \"nova-scheduler-0\" (UID: \"441334b1-15af-4e17-8974-c5ab19338eee\") " pod="openstack/nova-scheduler-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.265672 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441334b1-15af-4e17-8974-c5ab19338eee-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"441334b1-15af-4e17-8974-c5ab19338eee\") " pod="openstack/nova-scheduler-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.265707 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/441334b1-15af-4e17-8974-c5ab19338eee-config-data\") pod \"nova-scheduler-0\" (UID: \"441334b1-15af-4e17-8974-c5ab19338eee\") " pod="openstack/nova-scheduler-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.367611 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/441334b1-15af-4e17-8974-c5ab19338eee-config-data\") pod \"nova-scheduler-0\" (UID: \"441334b1-15af-4e17-8974-c5ab19338eee\") " pod="openstack/nova-scheduler-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.367815 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v8j7\" (UniqueName: \"kubernetes.io/projected/441334b1-15af-4e17-8974-c5ab19338eee-kube-api-access-8v8j7\") pod \"nova-scheduler-0\" (UID: \"441334b1-15af-4e17-8974-c5ab19338eee\") " pod="openstack/nova-scheduler-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.367900 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441334b1-15af-4e17-8974-c5ab19338eee-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"441334b1-15af-4e17-8974-c5ab19338eee\") " pod="openstack/nova-scheduler-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.372052 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441334b1-15af-4e17-8974-c5ab19338eee-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"441334b1-15af-4e17-8974-c5ab19338eee\") " pod="openstack/nova-scheduler-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.372211 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/441334b1-15af-4e17-8974-c5ab19338eee-config-data\") pod \"nova-scheduler-0\" (UID: \"441334b1-15af-4e17-8974-c5ab19338eee\") " pod="openstack/nova-scheduler-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.383672 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v8j7\" (UniqueName: \"kubernetes.io/projected/441334b1-15af-4e17-8974-c5ab19338eee-kube-api-access-8v8j7\") pod \"nova-scheduler-0\" (UID: \"441334b1-15af-4e17-8974-c5ab19338eee\") " pod="openstack/nova-scheduler-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.458826 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.476634 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.496938 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.498904 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.501657 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.501766 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.507188 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.508955 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.511158 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.573109 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-public-tls-certs\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.573206 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.573253 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhx4n\" (UniqueName: \"kubernetes.io/projected/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-kube-api-access-zhx4n\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.573275 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-config-data\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.573375 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-logs\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.573396 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.675246 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-public-tls-certs\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.675353 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.675396 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhx4n\" (UniqueName: \"kubernetes.io/projected/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-kube-api-access-zhx4n\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.675484 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-config-data\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.675826 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-logs\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.675884 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.676253 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-logs\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.680098 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-config-data\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.680727 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-public-tls-certs\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.681049 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.690858 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhx4n\" (UniqueName: \"kubernetes.io/projected/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-kube-api-access-zhx4n\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.692255 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2ffb182-594c-44ce-9d99-32ea22b5bfe9-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f2ffb182-594c-44ce-9d99-32ea22b5bfe9\") " pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.818758 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.952993 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86841bf5-3d74-480b-b7b4-ffb6534fb441" path="/var/lib/kubelet/pods/86841bf5-3d74-480b-b7b4-ffb6534fb441/volumes" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.953988 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9e0be7f-849b-4873-83fd-c0a440f5e9e3" path="/var/lib/kubelet/pods/c9e0be7f-849b-4873-83fd-c0a440f5e9e3/volumes" Feb 25 13:41:26 crc kubenswrapper[4815]: I0225 13:41:26.959049 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 13:41:26 crc kubenswrapper[4815]: W0225 13:41:26.964773 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod441334b1_15af_4e17_8974_c5ab19338eee.slice/crio-00190fb74ad85966e2552901745b733a6d68251e9ec7890dc0dc6c1196a3b6fa WatchSource:0}: Error finding container 00190fb74ad85966e2552901745b733a6d68251e9ec7890dc0dc6c1196a3b6fa: Status 404 returned error can't find the container with id 00190fb74ad85966e2552901745b733a6d68251e9ec7890dc0dc6c1196a3b6fa Feb 25 13:41:27 crc kubenswrapper[4815]: I0225 13:41:27.141986 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"441334b1-15af-4e17-8974-c5ab19338eee","Type":"ContainerStarted","Data":"f3bd8cb0a00d67565f474e0a2c0e844477197ee794c9093a5c1e89af3699ad5a"} Feb 25 13:41:27 crc kubenswrapper[4815]: I0225 13:41:27.142038 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"441334b1-15af-4e17-8974-c5ab19338eee","Type":"ContainerStarted","Data":"00190fb74ad85966e2552901745b733a6d68251e9ec7890dc0dc6c1196a3b6fa"} Feb 25 13:41:27 crc kubenswrapper[4815]: I0225 13:41:27.159177 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.159159547 podStartE2EDuration="1.159159547s" podCreationTimestamp="2026-02-25 13:41:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:41:27.156412593 +0000 UTC m=+1244.957510667" watchObservedRunningTime="2026-02-25 13:41:27.159159547 +0000 UTC m=+1244.960257601" Feb 25 13:41:27 crc kubenswrapper[4815]: W0225 13:41:27.261707 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2ffb182_594c_44ce_9d99_32ea22b5bfe9.slice/crio-ba0ea26a1fdb572d793ad03b08d54f24275b4b4a472519a61f631ff42011e559 WatchSource:0}: Error finding container ba0ea26a1fdb572d793ad03b08d54f24275b4b4a472519a61f631ff42011e559: Status 404 returned error can't find the container with id ba0ea26a1fdb572d793ad03b08d54f24275b4b4a472519a61f631ff42011e559 Feb 25 13:41:27 crc kubenswrapper[4815]: I0225 13:41:27.262796 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 13:41:28 crc kubenswrapper[4815]: I0225 13:41:28.157771 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f2ffb182-594c-44ce-9d99-32ea22b5bfe9","Type":"ContainerStarted","Data":"85d279359bb70db9b36c19118a04ff6952a966305e36c94a9b591d236b737ceb"} Feb 25 13:41:28 crc kubenswrapper[4815]: I0225 13:41:28.158254 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f2ffb182-594c-44ce-9d99-32ea22b5bfe9","Type":"ContainerStarted","Data":"a15345cf54ad19715b90cc1b76ed4793a94950ae3740b0c0f66debeee5ab91bd"} Feb 25 13:41:28 crc kubenswrapper[4815]: I0225 13:41:28.158280 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f2ffb182-594c-44ce-9d99-32ea22b5bfe9","Type":"ContainerStarted","Data":"ba0ea26a1fdb572d793ad03b08d54f24275b4b4a472519a61f631ff42011e559"} Feb 25 13:41:29 crc kubenswrapper[4815]: I0225 13:41:29.776979 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 25 13:41:29 crc kubenswrapper[4815]: I0225 13:41:29.777995 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 25 13:41:31 crc kubenswrapper[4815]: I0225 13:41:31.512189 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 25 13:41:34 crc kubenswrapper[4815]: I0225 13:41:34.778025 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 25 13:41:34 crc kubenswrapper[4815]: I0225 13:41:34.778539 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 25 13:41:35 crc kubenswrapper[4815]: I0225 13:41:35.794779 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d20eee08-f0d7-49bb-a195-202a600195f9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 13:41:35 crc kubenswrapper[4815]: I0225 13:41:35.794785 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d20eee08-f0d7-49bb-a195-202a600195f9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 13:41:36 crc kubenswrapper[4815]: I0225 13:41:36.512355 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 25 13:41:36 crc kubenswrapper[4815]: I0225 13:41:36.539426 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 25 13:41:36 crc kubenswrapper[4815]: I0225 13:41:36.569199 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=10.5691814 podStartE2EDuration="10.5691814s" podCreationTimestamp="2026-02-25 13:41:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:41:28.192797645 +0000 UTC m=+1245.993895739" watchObservedRunningTime="2026-02-25 13:41:36.5691814 +0000 UTC m=+1254.370279454" Feb 25 13:41:36 crc kubenswrapper[4815]: I0225 13:41:36.819955 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 13:41:36 crc kubenswrapper[4815]: I0225 13:41:36.820640 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 13:41:37 crc kubenswrapper[4815]: I0225 13:41:37.293391 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 25 13:41:37 crc kubenswrapper[4815]: I0225 13:41:37.832726 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f2ffb182-594c-44ce-9d99-32ea22b5bfe9" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.216:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 13:41:37 crc kubenswrapper[4815]: I0225 13:41:37.832726 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f2ffb182-594c-44ce-9d99-32ea22b5bfe9" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.216:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 13:41:39 crc kubenswrapper[4815]: I0225 13:41:39.229702 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 25 13:41:44 crc kubenswrapper[4815]: I0225 13:41:44.784921 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 25 13:41:44 crc kubenswrapper[4815]: I0225 13:41:44.790304 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 25 13:41:44 crc kubenswrapper[4815]: I0225 13:41:44.802013 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 25 13:41:45 crc kubenswrapper[4815]: I0225 13:41:45.371354 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 25 13:41:46 crc kubenswrapper[4815]: I0225 13:41:46.828052 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 25 13:41:46 crc kubenswrapper[4815]: I0225 13:41:46.831182 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 25 13:41:46 crc kubenswrapper[4815]: I0225 13:41:46.835456 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 25 13:41:46 crc kubenswrapper[4815]: I0225 13:41:46.841610 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 25 13:41:47 crc kubenswrapper[4815]: I0225 13:41:47.383068 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 25 13:41:47 crc kubenswrapper[4815]: I0225 13:41:47.393620 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 25 13:41:55 crc kubenswrapper[4815]: I0225 13:41:55.070406 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 13:41:56 crc kubenswrapper[4815]: I0225 13:41:56.256205 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 13:41:59 crc kubenswrapper[4815]: I0225 13:41:59.973838 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" containerName="rabbitmq" containerID="cri-o://7e7cb8ec442c6d0abc60fe0d32bc0d3e4bb7cbc24ea616784c6b8eba67f393c7" gracePeriod=604796 Feb 25 13:42:00 crc kubenswrapper[4815]: I0225 13:42:00.142943 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533782-cw687"] Feb 25 13:42:00 crc kubenswrapper[4815]: I0225 13:42:00.144401 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533782-cw687" Feb 25 13:42:00 crc kubenswrapper[4815]: I0225 13:42:00.147107 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:42:00 crc kubenswrapper[4815]: I0225 13:42:00.147660 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:42:00 crc kubenswrapper[4815]: I0225 13:42:00.148064 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:42:00 crc kubenswrapper[4815]: I0225 13:42:00.168157 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533782-cw687"] Feb 25 13:42:00 crc kubenswrapper[4815]: I0225 13:42:00.238490 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="0b98e7b2-3068-4f59-b25f-4755bd752b49" containerName="rabbitmq" containerID="cri-o://9abd009533a8d747d7e9365f84325c58794cd02960b609440095ab81603d616c" gracePeriod=604797 Feb 25 13:42:00 crc kubenswrapper[4815]: I0225 13:42:00.264178 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjcxg\" (UniqueName: \"kubernetes.io/projected/34838863-42b3-440d-96be-a1a0c48b32ed-kube-api-access-sjcxg\") pod \"auto-csr-approver-29533782-cw687\" (UID: \"34838863-42b3-440d-96be-a1a0c48b32ed\") " pod="openshift-infra/auto-csr-approver-29533782-cw687" Feb 25 13:42:00 crc kubenswrapper[4815]: I0225 13:42:00.365775 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjcxg\" (UniqueName: \"kubernetes.io/projected/34838863-42b3-440d-96be-a1a0c48b32ed-kube-api-access-sjcxg\") pod \"auto-csr-approver-29533782-cw687\" (UID: \"34838863-42b3-440d-96be-a1a0c48b32ed\") " pod="openshift-infra/auto-csr-approver-29533782-cw687" Feb 25 13:42:00 crc kubenswrapper[4815]: I0225 13:42:00.402715 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjcxg\" (UniqueName: \"kubernetes.io/projected/34838863-42b3-440d-96be-a1a0c48b32ed-kube-api-access-sjcxg\") pod \"auto-csr-approver-29533782-cw687\" (UID: \"34838863-42b3-440d-96be-a1a0c48b32ed\") " pod="openshift-infra/auto-csr-approver-29533782-cw687" Feb 25 13:42:00 crc kubenswrapper[4815]: I0225 13:42:00.473986 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533782-cw687" Feb 25 13:42:01 crc kubenswrapper[4815]: I0225 13:42:01.043043 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533782-cw687"] Feb 25 13:42:01 crc kubenswrapper[4815]: I0225 13:42:01.541529 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533782-cw687" event={"ID":"34838863-42b3-440d-96be-a1a0c48b32ed","Type":"ContainerStarted","Data":"294af14336a0f69d19ada87fad236973c5c28fdc4e62cf27680509edbeaadf8b"} Feb 25 13:42:02 crc kubenswrapper[4815]: I0225 13:42:02.552162 4815 generic.go:334] "Generic (PLEG): container finished" podID="34838863-42b3-440d-96be-a1a0c48b32ed" containerID="ebb6d812cafbfdf225df72faccc76a98ea8fc505a950399f7bcacf5169f65982" exitCode=0 Feb 25 13:42:02 crc kubenswrapper[4815]: I0225 13:42:02.552216 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533782-cw687" event={"ID":"34838863-42b3-440d-96be-a1a0c48b32ed","Type":"ContainerDied","Data":"ebb6d812cafbfdf225df72faccc76a98ea8fc505a950399f7bcacf5169f65982"} Feb 25 13:42:03 crc kubenswrapper[4815]: I0225 13:42:03.954936 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533782-cw687" Feb 25 13:42:04 crc kubenswrapper[4815]: I0225 13:42:04.040108 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjcxg\" (UniqueName: \"kubernetes.io/projected/34838863-42b3-440d-96be-a1a0c48b32ed-kube-api-access-sjcxg\") pod \"34838863-42b3-440d-96be-a1a0c48b32ed\" (UID: \"34838863-42b3-440d-96be-a1a0c48b32ed\") " Feb 25 13:42:04 crc kubenswrapper[4815]: I0225 13:42:04.049003 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34838863-42b3-440d-96be-a1a0c48b32ed-kube-api-access-sjcxg" (OuterVolumeSpecName: "kube-api-access-sjcxg") pod "34838863-42b3-440d-96be-a1a0c48b32ed" (UID: "34838863-42b3-440d-96be-a1a0c48b32ed"). InnerVolumeSpecName "kube-api-access-sjcxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:42:04 crc kubenswrapper[4815]: I0225 13:42:04.143071 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjcxg\" (UniqueName: \"kubernetes.io/projected/34838863-42b3-440d-96be-a1a0c48b32ed-kube-api-access-sjcxg\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:04 crc kubenswrapper[4815]: I0225 13:42:04.585774 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533782-cw687" event={"ID":"34838863-42b3-440d-96be-a1a0c48b32ed","Type":"ContainerDied","Data":"294af14336a0f69d19ada87fad236973c5c28fdc4e62cf27680509edbeaadf8b"} Feb 25 13:42:04 crc kubenswrapper[4815]: I0225 13:42:04.586033 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="294af14336a0f69d19ada87fad236973c5c28fdc4e62cf27680509edbeaadf8b" Feb 25 13:42:04 crc kubenswrapper[4815]: I0225 13:42:04.585899 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533782-cw687" Feb 25 13:42:05 crc kubenswrapper[4815]: I0225 13:42:05.042578 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533776-kz9qn"] Feb 25 13:42:05 crc kubenswrapper[4815]: I0225 13:42:05.053400 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533776-kz9qn"] Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.557490 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.589158 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-gs7cc"] Feb 25 13:42:06 crc kubenswrapper[4815]: E0225 13:42:06.589688 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" containerName="rabbitmq" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.589713 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" containerName="rabbitmq" Feb 25 13:42:06 crc kubenswrapper[4815]: E0225 13:42:06.589726 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" containerName="setup-container" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.589734 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" containerName="setup-container" Feb 25 13:42:06 crc kubenswrapper[4815]: E0225 13:42:06.589760 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34838863-42b3-440d-96be-a1a0c48b32ed" containerName="oc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.589768 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="34838863-42b3-440d-96be-a1a0c48b32ed" containerName="oc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.590006 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" containerName="rabbitmq" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.590047 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="34838863-42b3-440d-96be-a1a0c48b32ed" containerName="oc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.591261 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.594061 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.597588 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-erlang-cookie-secret\") pod \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.597716 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-pod-info\") pod \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.597745 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-plugins\") pod \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.597789 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-confd\") pod \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.597836 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-server-conf\") pod \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.597881 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-plugins-conf\") pod \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.597922 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-config-data\") pod \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.598006 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-erlang-cookie\") pod \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.598032 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.598050 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-tls\") pod \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.598077 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsfhb\" (UniqueName: \"kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-kube-api-access-zsfhb\") pod \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\" (UID: \"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.600179 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" (UID: "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.600567 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" (UID: "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.600908 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" (UID: "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.607753 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" (UID: "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.610849 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" (UID: "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.620772 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-kube-api-access-zsfhb" (OuterVolumeSpecName: "kube-api-access-zsfhb") pod "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" (UID: "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d"). InnerVolumeSpecName "kube-api-access-zsfhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.620782 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-pod-info" (OuterVolumeSpecName: "pod-info") pod "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" (UID: "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.620844 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" (UID: "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.641116 4815 generic.go:334] "Generic (PLEG): container finished" podID="4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" containerID="7e7cb8ec442c6d0abc60fe0d32bc0d3e4bb7cbc24ea616784c6b8eba67f393c7" exitCode=0 Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.641214 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d","Type":"ContainerDied","Data":"7e7cb8ec442c6d0abc60fe0d32bc0d3e4bb7cbc24ea616784c6b8eba67f393c7"} Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.641240 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4a91fa1d-2609-4b30-9ea1-8a5b10145c8d","Type":"ContainerDied","Data":"7949d637d954a9850bc8c2a36f253eb23f291f7be5db6df97664f033f9b4163f"} Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.641254 4815 scope.go:117] "RemoveContainer" containerID="7e7cb8ec442c6d0abc60fe0d32bc0d3e4bb7cbc24ea616784c6b8eba67f393c7" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.641383 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.683157 4815 generic.go:334] "Generic (PLEG): container finished" podID="0b98e7b2-3068-4f59-b25f-4755bd752b49" containerID="9abd009533a8d747d7e9365f84325c58794cd02960b609440095ab81603d616c" exitCode=0 Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.683224 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-gs7cc"] Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.683463 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0b98e7b2-3068-4f59-b25f-4755bd752b49","Type":"ContainerDied","Data":"9abd009533a8d747d7e9365f84325c58794cd02960b609440095ab81603d616c"} Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.711661 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.711966 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2z4w\" (UniqueName: \"kubernetes.io/projected/1b6433b4-ee14-4f19-8d64-e930f0ca0644-kube-api-access-f2z4w\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.712029 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.712076 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.712149 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-dns-svc\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.712371 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-config\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.712419 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.712750 4815 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.712774 4815 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-pod-info\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.712787 4815 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.712816 4815 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.712830 4815 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.712857 4815 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.712877 4815 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.712890 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsfhb\" (UniqueName: \"kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-kube-api-access-zsfhb\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.719027 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-config-data" (OuterVolumeSpecName: "config-data") pod "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" (UID: "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.743132 4815 scope.go:117] "RemoveContainer" containerID="e633733b61b772b04a5fcbcaa5076b5a7353c6b7fd094307d1bc64bbb9741d6c" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.763616 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-server-conf" (OuterVolumeSpecName: "server-conf") pod "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" (UID: "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.784963 4815 scope.go:117] "RemoveContainer" containerID="7e7cb8ec442c6d0abc60fe0d32bc0d3e4bb7cbc24ea616784c6b8eba67f393c7" Feb 25 13:42:06 crc kubenswrapper[4815]: E0225 13:42:06.785767 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e7cb8ec442c6d0abc60fe0d32bc0d3e4bb7cbc24ea616784c6b8eba67f393c7\": container with ID starting with 7e7cb8ec442c6d0abc60fe0d32bc0d3e4bb7cbc24ea616784c6b8eba67f393c7 not found: ID does not exist" containerID="7e7cb8ec442c6d0abc60fe0d32bc0d3e4bb7cbc24ea616784c6b8eba67f393c7" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.785814 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e7cb8ec442c6d0abc60fe0d32bc0d3e4bb7cbc24ea616784c6b8eba67f393c7"} err="failed to get container status \"7e7cb8ec442c6d0abc60fe0d32bc0d3e4bb7cbc24ea616784c6b8eba67f393c7\": rpc error: code = NotFound desc = could not find container \"7e7cb8ec442c6d0abc60fe0d32bc0d3e4bb7cbc24ea616784c6b8eba67f393c7\": container with ID starting with 7e7cb8ec442c6d0abc60fe0d32bc0d3e4bb7cbc24ea616784c6b8eba67f393c7 not found: ID does not exist" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.785845 4815 scope.go:117] "RemoveContainer" containerID="e633733b61b772b04a5fcbcaa5076b5a7353c6b7fd094307d1bc64bbb9741d6c" Feb 25 13:42:06 crc kubenswrapper[4815]: E0225 13:42:06.786595 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e633733b61b772b04a5fcbcaa5076b5a7353c6b7fd094307d1bc64bbb9741d6c\": container with ID starting with e633733b61b772b04a5fcbcaa5076b5a7353c6b7fd094307d1bc64bbb9741d6c not found: ID does not exist" containerID="e633733b61b772b04a5fcbcaa5076b5a7353c6b7fd094307d1bc64bbb9741d6c" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.786632 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e633733b61b772b04a5fcbcaa5076b5a7353c6b7fd094307d1bc64bbb9741d6c"} err="failed to get container status \"e633733b61b772b04a5fcbcaa5076b5a7353c6b7fd094307d1bc64bbb9741d6c\": rpc error: code = NotFound desc = could not find container \"e633733b61b772b04a5fcbcaa5076b5a7353c6b7fd094307d1bc64bbb9741d6c\": container with ID starting with e633733b61b772b04a5fcbcaa5076b5a7353c6b7fd094307d1bc64bbb9741d6c not found: ID does not exist" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.796701 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" (UID: "4a91fa1d-2609-4b30-9ea1-8a5b10145c8d"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.804723 4815 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.814858 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.814937 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.814982 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2z4w\" (UniqueName: \"kubernetes.io/projected/1b6433b4-ee14-4f19-8d64-e930f0ca0644-kube-api-access-f2z4w\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.815013 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.815042 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.815071 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-dns-svc\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.815110 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-config\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.815158 4815 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.815170 4815 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.815179 4815 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-server-conf\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.815187 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.816619 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-config\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.817654 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.817710 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-dns-svc\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.818319 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.818355 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.818497 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.832898 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.836440 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2z4w\" (UniqueName: \"kubernetes.io/projected/1b6433b4-ee14-4f19-8d64-e930f0ca0644-kube-api-access-f2z4w\") pod \"dnsmasq-dns-5576978c7c-gs7cc\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.916174 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-confd\") pod \"0b98e7b2-3068-4f59-b25f-4755bd752b49\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.916265 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-plugins\") pod \"0b98e7b2-3068-4f59-b25f-4755bd752b49\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.916293 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-plugins-conf\") pod \"0b98e7b2-3068-4f59-b25f-4755bd752b49\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.916327 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-server-conf\") pod \"0b98e7b2-3068-4f59-b25f-4755bd752b49\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.916364 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-erlang-cookie\") pod \"0b98e7b2-3068-4f59-b25f-4755bd752b49\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.916394 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-tls\") pod \"0b98e7b2-3068-4f59-b25f-4755bd752b49\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.916471 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-config-data\") pod \"0b98e7b2-3068-4f59-b25f-4755bd752b49\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.916539 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0b98e7b2-3068-4f59-b25f-4755bd752b49-erlang-cookie-secret\") pod \"0b98e7b2-3068-4f59-b25f-4755bd752b49\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.916572 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0b98e7b2-3068-4f59-b25f-4755bd752b49-pod-info\") pod \"0b98e7b2-3068-4f59-b25f-4755bd752b49\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.916591 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"0b98e7b2-3068-4f59-b25f-4755bd752b49\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.916623 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9gpj\" (UniqueName: \"kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-kube-api-access-h9gpj\") pod \"0b98e7b2-3068-4f59-b25f-4755bd752b49\" (UID: \"0b98e7b2-3068-4f59-b25f-4755bd752b49\") " Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.917956 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0b98e7b2-3068-4f59-b25f-4755bd752b49" (UID: "0b98e7b2-3068-4f59-b25f-4755bd752b49"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.919817 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0b98e7b2-3068-4f59-b25f-4755bd752b49" (UID: "0b98e7b2-3068-4f59-b25f-4755bd752b49"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.920053 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0b98e7b2-3068-4f59-b25f-4755bd752b49" (UID: "0b98e7b2-3068-4f59-b25f-4755bd752b49"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.924605 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0b98e7b2-3068-4f59-b25f-4755bd752b49-pod-info" (OuterVolumeSpecName: "pod-info") pod "0b98e7b2-3068-4f59-b25f-4755bd752b49" (UID: "0b98e7b2-3068-4f59-b25f-4755bd752b49"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.924915 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b98e7b2-3068-4f59-b25f-4755bd752b49-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0b98e7b2-3068-4f59-b25f-4755bd752b49" (UID: "0b98e7b2-3068-4f59-b25f-4755bd752b49"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.925454 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-kube-api-access-h9gpj" (OuterVolumeSpecName: "kube-api-access-h9gpj") pod "0b98e7b2-3068-4f59-b25f-4755bd752b49" (UID: "0b98e7b2-3068-4f59-b25f-4755bd752b49"). InnerVolumeSpecName "kube-api-access-h9gpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.925762 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "0b98e7b2-3068-4f59-b25f-4755bd752b49" (UID: "0b98e7b2-3068-4f59-b25f-4755bd752b49"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.928288 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0b98e7b2-3068-4f59-b25f-4755bd752b49" (UID: "0b98e7b2-3068-4f59-b25f-4755bd752b49"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.953486 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="805e511a-ead5-45ec-bcdc-018a16e7f059" path="/var/lib/kubelet/pods/805e511a-ead5-45ec-bcdc-018a16e7f059/volumes" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.967260 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-config-data" (OuterVolumeSpecName: "config-data") pod "0b98e7b2-3068-4f59-b25f-4755bd752b49" (UID: "0b98e7b2-3068-4f59-b25f-4755bd752b49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:42:06 crc kubenswrapper[4815]: I0225 13:42:06.989825 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-server-conf" (OuterVolumeSpecName: "server-conf") pod "0b98e7b2-3068-4f59-b25f-4755bd752b49" (UID: "0b98e7b2-3068-4f59-b25f-4755bd752b49"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.002680 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.013007 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.014441 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.020339 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.020379 4815 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0b98e7b2-3068-4f59-b25f-4755bd752b49-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.020394 4815 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0b98e7b2-3068-4f59-b25f-4755bd752b49-pod-info\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.020430 4815 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.020442 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9gpj\" (UniqueName: \"kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-kube-api-access-h9gpj\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.020453 4815 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.022528 4815 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.022553 4815 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0b98e7b2-3068-4f59-b25f-4755bd752b49-server-conf\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.022565 4815 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.022578 4815 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.022440 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 13:42:07 crc kubenswrapper[4815]: E0225 13:42:07.022983 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b98e7b2-3068-4f59-b25f-4755bd752b49" containerName="rabbitmq" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.023003 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b98e7b2-3068-4f59-b25f-4755bd752b49" containerName="rabbitmq" Feb 25 13:42:07 crc kubenswrapper[4815]: E0225 13:42:07.023033 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b98e7b2-3068-4f59-b25f-4755bd752b49" containerName="setup-container" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.023040 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b98e7b2-3068-4f59-b25f-4755bd752b49" containerName="setup-container" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.023214 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b98e7b2-3068-4f59-b25f-4755bd752b49" containerName="rabbitmq" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.024188 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.026964 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.027230 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.027446 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.027624 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.027783 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.029294 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.029529 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-2m84j" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.044963 4815 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.059410 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.074145 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0b98e7b2-3068-4f59-b25f-4755bd752b49" (UID: "0b98e7b2-3068-4f59-b25f-4755bd752b49"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.124380 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bd5b84a1-88ec-4727-9683-e696540ec165-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.124426 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bd5b84a1-88ec-4727-9683-e696540ec165-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.124452 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.124682 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bd5b84a1-88ec-4727-9683-e696540ec165-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.124800 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bd5b84a1-88ec-4727-9683-e696540ec165-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.124841 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bd5b84a1-88ec-4727-9683-e696540ec165-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.124889 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd5b84a1-88ec-4727-9683-e696540ec165-config-data\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.125021 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-727f5\" (UniqueName: \"kubernetes.io/projected/bd5b84a1-88ec-4727-9683-e696540ec165-kube-api-access-727f5\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.125125 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bd5b84a1-88ec-4727-9683-e696540ec165-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.125171 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bd5b84a1-88ec-4727-9683-e696540ec165-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.125219 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bd5b84a1-88ec-4727-9683-e696540ec165-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.125370 4815 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.125394 4815 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0b98e7b2-3068-4f59-b25f-4755bd752b49-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.226982 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bd5b84a1-88ec-4727-9683-e696540ec165-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.227036 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bd5b84a1-88ec-4727-9683-e696540ec165-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.227055 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bd5b84a1-88ec-4727-9683-e696540ec165-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.227071 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd5b84a1-88ec-4727-9683-e696540ec165-config-data\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.227116 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-727f5\" (UniqueName: \"kubernetes.io/projected/bd5b84a1-88ec-4727-9683-e696540ec165-kube-api-access-727f5\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.227141 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bd5b84a1-88ec-4727-9683-e696540ec165-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.227157 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bd5b84a1-88ec-4727-9683-e696540ec165-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.227176 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bd5b84a1-88ec-4727-9683-e696540ec165-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.227238 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bd5b84a1-88ec-4727-9683-e696540ec165-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.227260 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bd5b84a1-88ec-4727-9683-e696540ec165-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.227283 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.228744 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bd5b84a1-88ec-4727-9683-e696540ec165-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.228833 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd5b84a1-88ec-4727-9683-e696540ec165-config-data\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.228842 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.228912 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bd5b84a1-88ec-4727-9683-e696540ec165-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.229733 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bd5b84a1-88ec-4727-9683-e696540ec165-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.230290 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bd5b84a1-88ec-4727-9683-e696540ec165-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.241650 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bd5b84a1-88ec-4727-9683-e696540ec165-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.241840 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bd5b84a1-88ec-4727-9683-e696540ec165-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.244010 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bd5b84a1-88ec-4727-9683-e696540ec165-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.244064 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bd5b84a1-88ec-4727-9683-e696540ec165-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.249967 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-727f5\" (UniqueName: \"kubernetes.io/projected/bd5b84a1-88ec-4727-9683-e696540ec165-kube-api-access-727f5\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.282386 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"bd5b84a1-88ec-4727-9683-e696540ec165\") " pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.352143 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.461657 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-gs7cc"] Feb 25 13:42:07 crc kubenswrapper[4815]: W0225 13:42:07.466191 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b6433b4_ee14_4f19_8d64_e930f0ca0644.slice/crio-ca44b908a687eac82a3303635096875a56859fa89956b6c38971a332fab28552 WatchSource:0}: Error finding container ca44b908a687eac82a3303635096875a56859fa89956b6c38971a332fab28552: Status 404 returned error can't find the container with id ca44b908a687eac82a3303635096875a56859fa89956b6c38971a332fab28552 Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.711209 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0b98e7b2-3068-4f59-b25f-4755bd752b49","Type":"ContainerDied","Data":"69a9e634865d2724b24095ef53b1d262fd17ab671b5c78bd4ad7379533e7b58c"} Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.711554 4815 scope.go:117] "RemoveContainer" containerID="9abd009533a8d747d7e9365f84325c58794cd02960b609440095ab81603d616c" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.711694 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.717655 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" event={"ID":"1b6433b4-ee14-4f19-8d64-e930f0ca0644","Type":"ContainerStarted","Data":"ca44b908a687eac82a3303635096875a56859fa89956b6c38971a332fab28552"} Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.746278 4815 scope.go:117] "RemoveContainer" containerID="960d7fcb921daf08a33279bdf194bd3ec4837f2737f63bcd0cb1c55a13429919" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.756103 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.777875 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.786956 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.788883 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.796445 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.797171 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.797333 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.797442 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.797563 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.797992 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-js72s" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.798150 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.798151 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.846579 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1d3e5784-65f2-4282-a148-b4b015e5f875-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.846620 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1d3e5784-65f2-4282-a148-b4b015e5f875-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.846637 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1d3e5784-65f2-4282-a148-b4b015e5f875-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.846690 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqk84\" (UniqueName: \"kubernetes.io/projected/1d3e5784-65f2-4282-a148-b4b015e5f875-kube-api-access-jqk84\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.846723 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1d3e5784-65f2-4282-a148-b4b015e5f875-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.846740 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1d3e5784-65f2-4282-a148-b4b015e5f875-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.846816 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1d3e5784-65f2-4282-a148-b4b015e5f875-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.846838 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1d3e5784-65f2-4282-a148-b4b015e5f875-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.846859 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1d3e5784-65f2-4282-a148-b4b015e5f875-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.846924 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.846963 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1d3e5784-65f2-4282-a148-b4b015e5f875-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.850718 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.949072 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.949172 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1d3e5784-65f2-4282-a148-b4b015e5f875-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.949239 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1d3e5784-65f2-4282-a148-b4b015e5f875-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.949289 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1d3e5784-65f2-4282-a148-b4b015e5f875-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.949328 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1d3e5784-65f2-4282-a148-b4b015e5f875-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.949477 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqk84\" (UniqueName: \"kubernetes.io/projected/1d3e5784-65f2-4282-a148-b4b015e5f875-kube-api-access-jqk84\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.949576 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.949739 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1d3e5784-65f2-4282-a148-b4b015e5f875-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.951014 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1d3e5784-65f2-4282-a148-b4b015e5f875-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.954115 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1d3e5784-65f2-4282-a148-b4b015e5f875-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.954887 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1d3e5784-65f2-4282-a148-b4b015e5f875-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.963670 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1d3e5784-65f2-4282-a148-b4b015e5f875-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.963736 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1d3e5784-65f2-4282-a148-b4b015e5f875-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.963952 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1d3e5784-65f2-4282-a148-b4b015e5f875-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.963997 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1d3e5784-65f2-4282-a148-b4b015e5f875-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.964023 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1d3e5784-65f2-4282-a148-b4b015e5f875-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.964028 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1d3e5784-65f2-4282-a148-b4b015e5f875-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.964452 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1d3e5784-65f2-4282-a148-b4b015e5f875-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.965165 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1d3e5784-65f2-4282-a148-b4b015e5f875-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.969079 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1d3e5784-65f2-4282-a148-b4b015e5f875-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.969343 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1d3e5784-65f2-4282-a148-b4b015e5f875-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.970059 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqk84\" (UniqueName: \"kubernetes.io/projected/1d3e5784-65f2-4282-a148-b4b015e5f875-kube-api-access-jqk84\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:07 crc kubenswrapper[4815]: I0225 13:42:07.994636 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1d3e5784-65f2-4282-a148-b4b015e5f875\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:08 crc kubenswrapper[4815]: I0225 13:42:08.127253 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:08 crc kubenswrapper[4815]: I0225 13:42:08.621822 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 13:42:08 crc kubenswrapper[4815]: W0225 13:42:08.628377 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d3e5784_65f2_4282_a148_b4b015e5f875.slice/crio-fa46f9135f897b4f49fb0c6e8e40cbd1d1c9a4db04a8c6852ea59374bbc23303 WatchSource:0}: Error finding container fa46f9135f897b4f49fb0c6e8e40cbd1d1c9a4db04a8c6852ea59374bbc23303: Status 404 returned error can't find the container with id fa46f9135f897b4f49fb0c6e8e40cbd1d1c9a4db04a8c6852ea59374bbc23303 Feb 25 13:42:08 crc kubenswrapper[4815]: I0225 13:42:08.735109 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1d3e5784-65f2-4282-a148-b4b015e5f875","Type":"ContainerStarted","Data":"fa46f9135f897b4f49fb0c6e8e40cbd1d1c9a4db04a8c6852ea59374bbc23303"} Feb 25 13:42:08 crc kubenswrapper[4815]: I0225 13:42:08.736501 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bd5b84a1-88ec-4727-9683-e696540ec165","Type":"ContainerStarted","Data":"08d73fb1e20a02f98671efa6fad1ca2c6faa40ed5e8cf1d0c5c8fdc3b296ad35"} Feb 25 13:42:08 crc kubenswrapper[4815]: I0225 13:42:08.739439 4815 generic.go:334] "Generic (PLEG): container finished" podID="1b6433b4-ee14-4f19-8d64-e930f0ca0644" containerID="604fa926ed920118809ed02e682141bba54063a7fe88ed15dfe91e2d9ae36eea" exitCode=0 Feb 25 13:42:08 crc kubenswrapper[4815]: I0225 13:42:08.739483 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" event={"ID":"1b6433b4-ee14-4f19-8d64-e930f0ca0644","Type":"ContainerDied","Data":"604fa926ed920118809ed02e682141bba54063a7fe88ed15dfe91e2d9ae36eea"} Feb 25 13:42:08 crc kubenswrapper[4815]: I0225 13:42:08.948604 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b98e7b2-3068-4f59-b25f-4755bd752b49" path="/var/lib/kubelet/pods/0b98e7b2-3068-4f59-b25f-4755bd752b49/volumes" Feb 25 13:42:08 crc kubenswrapper[4815]: I0225 13:42:08.951825 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a91fa1d-2609-4b30-9ea1-8a5b10145c8d" path="/var/lib/kubelet/pods/4a91fa1d-2609-4b30-9ea1-8a5b10145c8d/volumes" Feb 25 13:42:09 crc kubenswrapper[4815]: I0225 13:42:09.749949 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bd5b84a1-88ec-4727-9683-e696540ec165","Type":"ContainerStarted","Data":"109c683f4184f2db1c92b43357a7b8dfb0c163409129be88277810521c8785d1"} Feb 25 13:42:09 crc kubenswrapper[4815]: I0225 13:42:09.752145 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" event={"ID":"1b6433b4-ee14-4f19-8d64-e930f0ca0644","Type":"ContainerStarted","Data":"6c5a8ab3a584f3bbd91c7975554bb6cc6682da1637abae9e912c502bd3eb38cf"} Feb 25 13:42:09 crc kubenswrapper[4815]: I0225 13:42:09.752352 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:09 crc kubenswrapper[4815]: I0225 13:42:09.818758 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" podStartSLOduration=3.818739403 podStartE2EDuration="3.818739403s" podCreationTimestamp="2026-02-25 13:42:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:42:09.801897828 +0000 UTC m=+1287.602995942" watchObservedRunningTime="2026-02-25 13:42:09.818739403 +0000 UTC m=+1287.619837457" Feb 25 13:42:10 crc kubenswrapper[4815]: I0225 13:42:10.761361 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1d3e5784-65f2-4282-a148-b4b015e5f875","Type":"ContainerStarted","Data":"c9da51c869fce82c4d0c29cc1c08d18894da464ae7bb71af132770c3c7d7b4e5"} Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.015736 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.100461 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-mzfxs"] Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.100841 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" podUID="292520ae-3dbf-4627-bf1d-b815d5af271b" containerName="dnsmasq-dns" containerID="cri-o://3f4ff5bb3070ffc987474d20227164b2c607635014c7ac60afc26ad6d0d677ea" gracePeriod=10 Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.286463 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-tft6z"] Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.288431 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.297752 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-tft6z"] Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.408974 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.409068 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.409159 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.409202 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ngvj\" (UniqueName: \"kubernetes.io/projected/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-kube-api-access-5ngvj\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.409226 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-config\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.409246 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.409284 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.511047 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.511422 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ngvj\" (UniqueName: \"kubernetes.io/projected/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-kube-api-access-5ngvj\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.511461 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-config\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.511490 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.511584 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.511616 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.511708 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.512653 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.512952 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.513060 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-config\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.513321 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.513328 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.513844 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.533982 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ngvj\" (UniqueName: \"kubernetes.io/projected/ece2c0c0-fcc9-4919-aa49-4b3df08531e4-kube-api-access-5ngvj\") pod \"dnsmasq-dns-8c6f6df99-tft6z\" (UID: \"ece2c0c0-fcc9-4919-aa49-4b3df08531e4\") " pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.633621 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.644324 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.816833 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-dns-svc\") pod \"292520ae-3dbf-4627-bf1d-b815d5af271b\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.818315 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-dns-swift-storage-0\") pod \"292520ae-3dbf-4627-bf1d-b815d5af271b\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.818389 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-ovsdbserver-sb\") pod \"292520ae-3dbf-4627-bf1d-b815d5af271b\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.818447 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dskch\" (UniqueName: \"kubernetes.io/projected/292520ae-3dbf-4627-bf1d-b815d5af271b-kube-api-access-dskch\") pod \"292520ae-3dbf-4627-bf1d-b815d5af271b\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.818472 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-ovsdbserver-nb\") pod \"292520ae-3dbf-4627-bf1d-b815d5af271b\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.818580 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-config\") pod \"292520ae-3dbf-4627-bf1d-b815d5af271b\" (UID: \"292520ae-3dbf-4627-bf1d-b815d5af271b\") " Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.824767 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/292520ae-3dbf-4627-bf1d-b815d5af271b-kube-api-access-dskch" (OuterVolumeSpecName: "kube-api-access-dskch") pod "292520ae-3dbf-4627-bf1d-b815d5af271b" (UID: "292520ae-3dbf-4627-bf1d-b815d5af271b"). InnerVolumeSpecName "kube-api-access-dskch". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.865473 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "292520ae-3dbf-4627-bf1d-b815d5af271b" (UID: "292520ae-3dbf-4627-bf1d-b815d5af271b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.876286 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "292520ae-3dbf-4627-bf1d-b815d5af271b" (UID: "292520ae-3dbf-4627-bf1d-b815d5af271b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.880132 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "292520ae-3dbf-4627-bf1d-b815d5af271b" (UID: "292520ae-3dbf-4627-bf1d-b815d5af271b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.881373 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-config" (OuterVolumeSpecName: "config") pod "292520ae-3dbf-4627-bf1d-b815d5af271b" (UID: "292520ae-3dbf-4627-bf1d-b815d5af271b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.885087 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "292520ae-3dbf-4627-bf1d-b815d5af271b" (UID: "292520ae-3dbf-4627-bf1d-b815d5af271b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.921999 4815 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.922038 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.922049 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dskch\" (UniqueName: \"kubernetes.io/projected/292520ae-3dbf-4627-bf1d-b815d5af271b-kube-api-access-dskch\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.922058 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.922070 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:17 crc kubenswrapper[4815]: I0225 13:42:17.922079 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292520ae-3dbf-4627-bf1d-b815d5af271b-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:18 crc kubenswrapper[4815]: I0225 13:42:18.090762 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-tft6z"] Feb 25 13:42:18 crc kubenswrapper[4815]: I0225 13:42:18.200300 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" event={"ID":"ece2c0c0-fcc9-4919-aa49-4b3df08531e4","Type":"ContainerStarted","Data":"ae598a13dd86015786f5a50253a3887054d8dbf6a058654dc731662409636cd1"} Feb 25 13:42:18 crc kubenswrapper[4815]: I0225 13:42:18.204990 4815 generic.go:334] "Generic (PLEG): container finished" podID="292520ae-3dbf-4627-bf1d-b815d5af271b" containerID="3f4ff5bb3070ffc987474d20227164b2c607635014c7ac60afc26ad6d0d677ea" exitCode=0 Feb 25 13:42:18 crc kubenswrapper[4815]: I0225 13:42:18.205171 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" event={"ID":"292520ae-3dbf-4627-bf1d-b815d5af271b","Type":"ContainerDied","Data":"3f4ff5bb3070ffc987474d20227164b2c607635014c7ac60afc26ad6d0d677ea"} Feb 25 13:42:18 crc kubenswrapper[4815]: I0225 13:42:18.205264 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" event={"ID":"292520ae-3dbf-4627-bf1d-b815d5af271b","Type":"ContainerDied","Data":"38af34e7109228ad9e0196a315f45e433a0b2af650168afb4240537996b210fd"} Feb 25 13:42:18 crc kubenswrapper[4815]: I0225 13:42:18.205342 4815 scope.go:117] "RemoveContainer" containerID="3f4ff5bb3070ffc987474d20227164b2c607635014c7ac60afc26ad6d0d677ea" Feb 25 13:42:18 crc kubenswrapper[4815]: I0225 13:42:18.205581 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-mzfxs" Feb 25 13:42:18 crc kubenswrapper[4815]: I0225 13:42:18.309619 4815 scope.go:117] "RemoveContainer" containerID="37e0b33154bb461902a98e3c518566038390dafbcafab1ed57ea6babbc26aa7f" Feb 25 13:42:18 crc kubenswrapper[4815]: I0225 13:42:18.337540 4815 scope.go:117] "RemoveContainer" containerID="3f4ff5bb3070ffc987474d20227164b2c607635014c7ac60afc26ad6d0d677ea" Feb 25 13:42:18 crc kubenswrapper[4815]: E0225 13:42:18.337941 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f4ff5bb3070ffc987474d20227164b2c607635014c7ac60afc26ad6d0d677ea\": container with ID starting with 3f4ff5bb3070ffc987474d20227164b2c607635014c7ac60afc26ad6d0d677ea not found: ID does not exist" containerID="3f4ff5bb3070ffc987474d20227164b2c607635014c7ac60afc26ad6d0d677ea" Feb 25 13:42:18 crc kubenswrapper[4815]: I0225 13:42:18.337985 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f4ff5bb3070ffc987474d20227164b2c607635014c7ac60afc26ad6d0d677ea"} err="failed to get container status \"3f4ff5bb3070ffc987474d20227164b2c607635014c7ac60afc26ad6d0d677ea\": rpc error: code = NotFound desc = could not find container \"3f4ff5bb3070ffc987474d20227164b2c607635014c7ac60afc26ad6d0d677ea\": container with ID starting with 3f4ff5bb3070ffc987474d20227164b2c607635014c7ac60afc26ad6d0d677ea not found: ID does not exist" Feb 25 13:42:18 crc kubenswrapper[4815]: I0225 13:42:18.338013 4815 scope.go:117] "RemoveContainer" containerID="37e0b33154bb461902a98e3c518566038390dafbcafab1ed57ea6babbc26aa7f" Feb 25 13:42:18 crc kubenswrapper[4815]: E0225 13:42:18.338467 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37e0b33154bb461902a98e3c518566038390dafbcafab1ed57ea6babbc26aa7f\": container with ID starting with 37e0b33154bb461902a98e3c518566038390dafbcafab1ed57ea6babbc26aa7f not found: ID does not exist" containerID="37e0b33154bb461902a98e3c518566038390dafbcafab1ed57ea6babbc26aa7f" Feb 25 13:42:18 crc kubenswrapper[4815]: I0225 13:42:18.338605 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37e0b33154bb461902a98e3c518566038390dafbcafab1ed57ea6babbc26aa7f"} err="failed to get container status \"37e0b33154bb461902a98e3c518566038390dafbcafab1ed57ea6babbc26aa7f\": rpc error: code = NotFound desc = could not find container \"37e0b33154bb461902a98e3c518566038390dafbcafab1ed57ea6babbc26aa7f\": container with ID starting with 37e0b33154bb461902a98e3c518566038390dafbcafab1ed57ea6babbc26aa7f not found: ID does not exist" Feb 25 13:42:18 crc kubenswrapper[4815]: I0225 13:42:18.342353 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-mzfxs"] Feb 25 13:42:18 crc kubenswrapper[4815]: I0225 13:42:18.351854 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-mzfxs"] Feb 25 13:42:18 crc kubenswrapper[4815]: I0225 13:42:18.957797 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="292520ae-3dbf-4627-bf1d-b815d5af271b" path="/var/lib/kubelet/pods/292520ae-3dbf-4627-bf1d-b815d5af271b/volumes" Feb 25 13:42:19 crc kubenswrapper[4815]: I0225 13:42:19.218252 4815 generic.go:334] "Generic (PLEG): container finished" podID="ece2c0c0-fcc9-4919-aa49-4b3df08531e4" containerID="5cedea2b790bad9f46261a2ed8a5bfe45adc74b17a879c229a09b28347d71bea" exitCode=0 Feb 25 13:42:19 crc kubenswrapper[4815]: I0225 13:42:19.218339 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" event={"ID":"ece2c0c0-fcc9-4919-aa49-4b3df08531e4","Type":"ContainerDied","Data":"5cedea2b790bad9f46261a2ed8a5bfe45adc74b17a879c229a09b28347d71bea"} Feb 25 13:42:20 crc kubenswrapper[4815]: I0225 13:42:20.232447 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" event={"ID":"ece2c0c0-fcc9-4919-aa49-4b3df08531e4","Type":"ContainerStarted","Data":"b05127266772226b1f132e431acabbbd8141f7cd82440e1a9f9527df5aa89a2f"} Feb 25 13:42:20 crc kubenswrapper[4815]: I0225 13:42:20.232712 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:20 crc kubenswrapper[4815]: I0225 13:42:20.262646 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" podStartSLOduration=3.262626989 podStartE2EDuration="3.262626989s" podCreationTimestamp="2026-02-25 13:42:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:42:20.256029508 +0000 UTC m=+1298.057127572" watchObservedRunningTime="2026-02-25 13:42:20.262626989 +0000 UTC m=+1298.063725063" Feb 25 13:42:21 crc kubenswrapper[4815]: I0225 13:42:21.342324 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:42:21 crc kubenswrapper[4815]: I0225 13:42:21.342669 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:42:27 crc kubenswrapper[4815]: I0225 13:42:27.646768 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8c6f6df99-tft6z" Feb 25 13:42:27 crc kubenswrapper[4815]: I0225 13:42:27.731142 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-gs7cc"] Feb 25 13:42:27 crc kubenswrapper[4815]: I0225 13:42:27.731680 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" podUID="1b6433b4-ee14-4f19-8d64-e930f0ca0644" containerName="dnsmasq-dns" containerID="cri-o://6c5a8ab3a584f3bbd91c7975554bb6cc6682da1637abae9e912c502bd3eb38cf" gracePeriod=10 Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.223542 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.318054 4815 generic.go:334] "Generic (PLEG): container finished" podID="1b6433b4-ee14-4f19-8d64-e930f0ca0644" containerID="6c5a8ab3a584f3bbd91c7975554bb6cc6682da1637abae9e912c502bd3eb38cf" exitCode=0 Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.318104 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" event={"ID":"1b6433b4-ee14-4f19-8d64-e930f0ca0644","Type":"ContainerDied","Data":"6c5a8ab3a584f3bbd91c7975554bb6cc6682da1637abae9e912c502bd3eb38cf"} Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.318140 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" event={"ID":"1b6433b4-ee14-4f19-8d64-e930f0ca0644","Type":"ContainerDied","Data":"ca44b908a687eac82a3303635096875a56859fa89956b6c38971a332fab28552"} Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.318164 4815 scope.go:117] "RemoveContainer" containerID="6c5a8ab3a584f3bbd91c7975554bb6cc6682da1637abae9e912c502bd3eb38cf" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.318316 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-gs7cc" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.335873 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-dns-swift-storage-0\") pod \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.335921 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-config\") pod \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.335965 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-ovsdbserver-nb\") pod \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.336063 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2z4w\" (UniqueName: \"kubernetes.io/projected/1b6433b4-ee14-4f19-8d64-e930f0ca0644-kube-api-access-f2z4w\") pod \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.336191 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-openstack-edpm-ipam\") pod \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.336333 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-dns-svc\") pod \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.336379 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-ovsdbserver-sb\") pod \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\" (UID: \"1b6433b4-ee14-4f19-8d64-e930f0ca0644\") " Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.345448 4815 scope.go:117] "RemoveContainer" containerID="604fa926ed920118809ed02e682141bba54063a7fe88ed15dfe91e2d9ae36eea" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.356726 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b6433b4-ee14-4f19-8d64-e930f0ca0644-kube-api-access-f2z4w" (OuterVolumeSpecName: "kube-api-access-f2z4w") pod "1b6433b4-ee14-4f19-8d64-e930f0ca0644" (UID: "1b6433b4-ee14-4f19-8d64-e930f0ca0644"). InnerVolumeSpecName "kube-api-access-f2z4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.394091 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-config" (OuterVolumeSpecName: "config") pod "1b6433b4-ee14-4f19-8d64-e930f0ca0644" (UID: "1b6433b4-ee14-4f19-8d64-e930f0ca0644"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.397742 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1b6433b4-ee14-4f19-8d64-e930f0ca0644" (UID: "1b6433b4-ee14-4f19-8d64-e930f0ca0644"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.407738 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1b6433b4-ee14-4f19-8d64-e930f0ca0644" (UID: "1b6433b4-ee14-4f19-8d64-e930f0ca0644"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.420425 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "1b6433b4-ee14-4f19-8d64-e930f0ca0644" (UID: "1b6433b4-ee14-4f19-8d64-e930f0ca0644"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.439160 4815 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.439205 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-config\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.439218 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.439229 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2z4w\" (UniqueName: \"kubernetes.io/projected/1b6433b4-ee14-4f19-8d64-e930f0ca0644-kube-api-access-f2z4w\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.439244 4815 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.441939 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1b6433b4-ee14-4f19-8d64-e930f0ca0644" (UID: "1b6433b4-ee14-4f19-8d64-e930f0ca0644"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.446156 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1b6433b4-ee14-4f19-8d64-e930f0ca0644" (UID: "1b6433b4-ee14-4f19-8d64-e930f0ca0644"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.480048 4815 scope.go:117] "RemoveContainer" containerID="6c5a8ab3a584f3bbd91c7975554bb6cc6682da1637abae9e912c502bd3eb38cf" Feb 25 13:42:28 crc kubenswrapper[4815]: E0225 13:42:28.480556 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c5a8ab3a584f3bbd91c7975554bb6cc6682da1637abae9e912c502bd3eb38cf\": container with ID starting with 6c5a8ab3a584f3bbd91c7975554bb6cc6682da1637abae9e912c502bd3eb38cf not found: ID does not exist" containerID="6c5a8ab3a584f3bbd91c7975554bb6cc6682da1637abae9e912c502bd3eb38cf" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.480633 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c5a8ab3a584f3bbd91c7975554bb6cc6682da1637abae9e912c502bd3eb38cf"} err="failed to get container status \"6c5a8ab3a584f3bbd91c7975554bb6cc6682da1637abae9e912c502bd3eb38cf\": rpc error: code = NotFound desc = could not find container \"6c5a8ab3a584f3bbd91c7975554bb6cc6682da1637abae9e912c502bd3eb38cf\": container with ID starting with 6c5a8ab3a584f3bbd91c7975554bb6cc6682da1637abae9e912c502bd3eb38cf not found: ID does not exist" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.480694 4815 scope.go:117] "RemoveContainer" containerID="604fa926ed920118809ed02e682141bba54063a7fe88ed15dfe91e2d9ae36eea" Feb 25 13:42:28 crc kubenswrapper[4815]: E0225 13:42:28.481108 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"604fa926ed920118809ed02e682141bba54063a7fe88ed15dfe91e2d9ae36eea\": container with ID starting with 604fa926ed920118809ed02e682141bba54063a7fe88ed15dfe91e2d9ae36eea not found: ID does not exist" containerID="604fa926ed920118809ed02e682141bba54063a7fe88ed15dfe91e2d9ae36eea" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.481144 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"604fa926ed920118809ed02e682141bba54063a7fe88ed15dfe91e2d9ae36eea"} err="failed to get container status \"604fa926ed920118809ed02e682141bba54063a7fe88ed15dfe91e2d9ae36eea\": rpc error: code = NotFound desc = could not find container \"604fa926ed920118809ed02e682141bba54063a7fe88ed15dfe91e2d9ae36eea\": container with ID starting with 604fa926ed920118809ed02e682141bba54063a7fe88ed15dfe91e2d9ae36eea not found: ID does not exist" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.541277 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.541316 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b6433b4-ee14-4f19-8d64-e930f0ca0644-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.661613 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-gs7cc"] Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.673553 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-gs7cc"] Feb 25 13:42:28 crc kubenswrapper[4815]: I0225 13:42:28.951411 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b6433b4-ee14-4f19-8d64-e930f0ca0644" path="/var/lib/kubelet/pods/1b6433b4-ee14-4f19-8d64-e930f0ca0644/volumes" Feb 25 13:42:39 crc kubenswrapper[4815]: I0225 13:42:39.784319 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j"] Feb 25 13:42:39 crc kubenswrapper[4815]: E0225 13:42:39.785382 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="292520ae-3dbf-4627-bf1d-b815d5af271b" containerName="init" Feb 25 13:42:39 crc kubenswrapper[4815]: I0225 13:42:39.785401 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="292520ae-3dbf-4627-bf1d-b815d5af271b" containerName="init" Feb 25 13:42:39 crc kubenswrapper[4815]: E0225 13:42:39.785426 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="292520ae-3dbf-4627-bf1d-b815d5af271b" containerName="dnsmasq-dns" Feb 25 13:42:39 crc kubenswrapper[4815]: I0225 13:42:39.785434 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="292520ae-3dbf-4627-bf1d-b815d5af271b" containerName="dnsmasq-dns" Feb 25 13:42:39 crc kubenswrapper[4815]: E0225 13:42:39.785460 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b6433b4-ee14-4f19-8d64-e930f0ca0644" containerName="init" Feb 25 13:42:39 crc kubenswrapper[4815]: I0225 13:42:39.785469 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b6433b4-ee14-4f19-8d64-e930f0ca0644" containerName="init" Feb 25 13:42:39 crc kubenswrapper[4815]: E0225 13:42:39.785496 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b6433b4-ee14-4f19-8d64-e930f0ca0644" containerName="dnsmasq-dns" Feb 25 13:42:39 crc kubenswrapper[4815]: I0225 13:42:39.785504 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b6433b4-ee14-4f19-8d64-e930f0ca0644" containerName="dnsmasq-dns" Feb 25 13:42:39 crc kubenswrapper[4815]: I0225 13:42:39.785931 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b6433b4-ee14-4f19-8d64-e930f0ca0644" containerName="dnsmasq-dns" Feb 25 13:42:39 crc kubenswrapper[4815]: I0225 13:42:39.785964 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="292520ae-3dbf-4627-bf1d-b815d5af271b" containerName="dnsmasq-dns" Feb 25 13:42:39 crc kubenswrapper[4815]: I0225 13:42:39.786591 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" Feb 25 13:42:39 crc kubenswrapper[4815]: I0225 13:42:39.788999 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 25 13:42:39 crc kubenswrapper[4815]: I0225 13:42:39.789001 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:42:39 crc kubenswrapper[4815]: I0225 13:42:39.789314 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 25 13:42:39 crc kubenswrapper[4815]: I0225 13:42:39.800102 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mmbj" Feb 25 13:42:39 crc kubenswrapper[4815]: I0225 13:42:39.802457 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j"] Feb 25 13:42:39 crc kubenswrapper[4815]: I0225 13:42:39.969088 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j\" (UID: \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" Feb 25 13:42:39 crc kubenswrapper[4815]: I0225 13:42:39.969357 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j\" (UID: \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" Feb 25 13:42:39 crc kubenswrapper[4815]: I0225 13:42:39.969648 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j\" (UID: \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" Feb 25 13:42:39 crc kubenswrapper[4815]: I0225 13:42:39.969854 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df8vx\" (UniqueName: \"kubernetes.io/projected/5bcbebb7-6865-44e4-b4da-61c06daeb9de-kube-api-access-df8vx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j\" (UID: \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" Feb 25 13:42:40 crc kubenswrapper[4815]: I0225 13:42:40.073546 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j\" (UID: \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" Feb 25 13:42:40 crc kubenswrapper[4815]: I0225 13:42:40.074641 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j\" (UID: \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" Feb 25 13:42:40 crc kubenswrapper[4815]: I0225 13:42:40.075229 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df8vx\" (UniqueName: \"kubernetes.io/projected/5bcbebb7-6865-44e4-b4da-61c06daeb9de-kube-api-access-df8vx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j\" (UID: \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" Feb 25 13:42:40 crc kubenswrapper[4815]: I0225 13:42:40.075396 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j\" (UID: \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" Feb 25 13:42:40 crc kubenswrapper[4815]: I0225 13:42:40.086001 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j\" (UID: \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" Feb 25 13:42:40 crc kubenswrapper[4815]: I0225 13:42:40.088941 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j\" (UID: \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" Feb 25 13:42:40 crc kubenswrapper[4815]: I0225 13:42:40.093943 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j\" (UID: \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" Feb 25 13:42:40 crc kubenswrapper[4815]: I0225 13:42:40.100585 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df8vx\" (UniqueName: \"kubernetes.io/projected/5bcbebb7-6865-44e4-b4da-61c06daeb9de-kube-api-access-df8vx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j\" (UID: \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" Feb 25 13:42:40 crc kubenswrapper[4815]: I0225 13:42:40.121917 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" Feb 25 13:42:40 crc kubenswrapper[4815]: I0225 13:42:40.720180 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j"] Feb 25 13:42:41 crc kubenswrapper[4815]: I0225 13:42:41.460788 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" event={"ID":"5bcbebb7-6865-44e4-b4da-61c06daeb9de","Type":"ContainerStarted","Data":"7df5ebdd5a6bd39d65470308c70f5dbc2a466a6284f08f25c3be3aeddacd0499"} Feb 25 13:42:42 crc kubenswrapper[4815]: I0225 13:42:42.473719 4815 generic.go:334] "Generic (PLEG): container finished" podID="bd5b84a1-88ec-4727-9683-e696540ec165" containerID="109c683f4184f2db1c92b43357a7b8dfb0c163409129be88277810521c8785d1" exitCode=0 Feb 25 13:42:42 crc kubenswrapper[4815]: I0225 13:42:42.473817 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bd5b84a1-88ec-4727-9683-e696540ec165","Type":"ContainerDied","Data":"109c683f4184f2db1c92b43357a7b8dfb0c163409129be88277810521c8785d1"} Feb 25 13:42:44 crc kubenswrapper[4815]: I0225 13:42:44.498320 4815 generic.go:334] "Generic (PLEG): container finished" podID="1d3e5784-65f2-4282-a148-b4b015e5f875" containerID="c9da51c869fce82c4d0c29cc1c08d18894da464ae7bb71af132770c3c7d7b4e5" exitCode=0 Feb 25 13:42:44 crc kubenswrapper[4815]: I0225 13:42:44.498407 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1d3e5784-65f2-4282-a148-b4b015e5f875","Type":"ContainerDied","Data":"c9da51c869fce82c4d0c29cc1c08d18894da464ae7bb71af132770c3c7d7b4e5"} Feb 25 13:42:44 crc kubenswrapper[4815]: I0225 13:42:44.502791 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bd5b84a1-88ec-4727-9683-e696540ec165","Type":"ContainerStarted","Data":"ff8f7d788060923ebae7cfdeff3927b6fb47dc504fc50907a6cbf40d8e0a7016"} Feb 25 13:42:44 crc kubenswrapper[4815]: I0225 13:42:44.503007 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 25 13:42:44 crc kubenswrapper[4815]: I0225 13:42:44.594612 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.594590129 podStartE2EDuration="38.594590129s" podCreationTimestamp="2026-02-25 13:42:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:42:44.581675596 +0000 UTC m=+1322.382773650" watchObservedRunningTime="2026-02-25 13:42:44.594590129 +0000 UTC m=+1322.395688193" Feb 25 13:42:45 crc kubenswrapper[4815]: I0225 13:42:45.240991 4815 scope.go:117] "RemoveContainer" containerID="555b0db9fc48103d6923e1891f56a10d3ab01f58a617ce1367f71c3820cfa3b0" Feb 25 13:42:46 crc kubenswrapper[4815]: I0225 13:42:46.533583 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1d3e5784-65f2-4282-a148-b4b015e5f875","Type":"ContainerStarted","Data":"10bd49c256120a740e627b44994a6f5e8e582acb43a24abbb702781b3e923ac8"} Feb 25 13:42:46 crc kubenswrapper[4815]: I0225 13:42:46.534377 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:42:46 crc kubenswrapper[4815]: I0225 13:42:46.558995 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=39.558977629 podStartE2EDuration="39.558977629s" podCreationTimestamp="2026-02-25 13:42:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 13:42:46.555945266 +0000 UTC m=+1324.357043320" watchObservedRunningTime="2026-02-25 13:42:46.558977629 +0000 UTC m=+1324.360075683" Feb 25 13:42:51 crc kubenswrapper[4815]: I0225 13:42:51.358266 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:42:51 crc kubenswrapper[4815]: I0225 13:42:51.358865 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:42:53 crc kubenswrapper[4815]: I0225 13:42:53.696596 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:42:54 crc kubenswrapper[4815]: I0225 13:42:54.607434 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" event={"ID":"5bcbebb7-6865-44e4-b4da-61c06daeb9de","Type":"ContainerStarted","Data":"c083b4f780a1daf0ddd856da845968d4e0a8dc8fbc6ea8fd167aee45774d8874"} Feb 25 13:42:54 crc kubenswrapper[4815]: I0225 13:42:54.634441 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" podStartSLOduration=2.6595236460000002 podStartE2EDuration="15.634416379s" podCreationTimestamp="2026-02-25 13:42:39 +0000 UTC" firstStartedPulling="2026-02-25 13:42:40.719420793 +0000 UTC m=+1318.520518847" lastFinishedPulling="2026-02-25 13:42:53.694313526 +0000 UTC m=+1331.495411580" observedRunningTime="2026-02-25 13:42:54.631291504 +0000 UTC m=+1332.432389598" watchObservedRunningTime="2026-02-25 13:42:54.634416379 +0000 UTC m=+1332.435514473" Feb 25 13:42:57 crc kubenswrapper[4815]: I0225 13:42:57.358921 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 25 13:42:58 crc kubenswrapper[4815]: I0225 13:42:58.135374 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 25 13:43:05 crc kubenswrapper[4815]: I0225 13:43:05.727564 4815 generic.go:334] "Generic (PLEG): container finished" podID="5bcbebb7-6865-44e4-b4da-61c06daeb9de" containerID="c083b4f780a1daf0ddd856da845968d4e0a8dc8fbc6ea8fd167aee45774d8874" exitCode=0 Feb 25 13:43:05 crc kubenswrapper[4815]: I0225 13:43:05.727667 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" event={"ID":"5bcbebb7-6865-44e4-b4da-61c06daeb9de","Type":"ContainerDied","Data":"c083b4f780a1daf0ddd856da845968d4e0a8dc8fbc6ea8fd167aee45774d8874"} Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.166574 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.240324 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-repo-setup-combined-ca-bundle\") pod \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\" (UID: \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\") " Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.240392 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-ssh-key-openstack-edpm-ipam\") pod \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\" (UID: \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\") " Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.240437 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-inventory\") pod \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\" (UID: \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\") " Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.240487 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df8vx\" (UniqueName: \"kubernetes.io/projected/5bcbebb7-6865-44e4-b4da-61c06daeb9de-kube-api-access-df8vx\") pod \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\" (UID: \"5bcbebb7-6865-44e4-b4da-61c06daeb9de\") " Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.251307 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "5bcbebb7-6865-44e4-b4da-61c06daeb9de" (UID: "5bcbebb7-6865-44e4-b4da-61c06daeb9de"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.253456 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bcbebb7-6865-44e4-b4da-61c06daeb9de-kube-api-access-df8vx" (OuterVolumeSpecName: "kube-api-access-df8vx") pod "5bcbebb7-6865-44e4-b4da-61c06daeb9de" (UID: "5bcbebb7-6865-44e4-b4da-61c06daeb9de"). InnerVolumeSpecName "kube-api-access-df8vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.273485 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-inventory" (OuterVolumeSpecName: "inventory") pod "5bcbebb7-6865-44e4-b4da-61c06daeb9de" (UID: "5bcbebb7-6865-44e4-b4da-61c06daeb9de"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.275306 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5bcbebb7-6865-44e4-b4da-61c06daeb9de" (UID: "5bcbebb7-6865-44e4-b4da-61c06daeb9de"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.343492 4815 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.343558 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.343581 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5bcbebb7-6865-44e4-b4da-61c06daeb9de-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.343599 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df8vx\" (UniqueName: \"kubernetes.io/projected/5bcbebb7-6865-44e4-b4da-61c06daeb9de-kube-api-access-df8vx\") on node \"crc\" DevicePath \"\"" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.747924 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" event={"ID":"5bcbebb7-6865-44e4-b4da-61c06daeb9de","Type":"ContainerDied","Data":"7df5ebdd5a6bd39d65470308c70f5dbc2a466a6284f08f25c3be3aeddacd0499"} Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.747968 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7df5ebdd5a6bd39d65470308c70f5dbc2a466a6284f08f25c3be3aeddacd0499" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.748026 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.860603 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr"] Feb 25 13:43:07 crc kubenswrapper[4815]: E0225 13:43:07.861315 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bcbebb7-6865-44e4-b4da-61c06daeb9de" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.861359 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bcbebb7-6865-44e4-b4da-61c06daeb9de" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.861829 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bcbebb7-6865-44e4-b4da-61c06daeb9de" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.863108 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.865365 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.865439 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.865874 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mmbj" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.867440 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.869590 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr"] Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.971608 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/017b7720-a334-4276-9531-a49fb9558275-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-t8psr\" (UID: \"017b7720-a334-4276-9531-a49fb9558275\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.972340 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/017b7720-a334-4276-9531-a49fb9558275-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-t8psr\" (UID: \"017b7720-a334-4276-9531-a49fb9558275\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" Feb 25 13:43:07 crc kubenswrapper[4815]: I0225 13:43:07.972384 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4frrm\" (UniqueName: \"kubernetes.io/projected/017b7720-a334-4276-9531-a49fb9558275-kube-api-access-4frrm\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-t8psr\" (UID: \"017b7720-a334-4276-9531-a49fb9558275\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" Feb 25 13:43:08 crc kubenswrapper[4815]: I0225 13:43:08.076577 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/017b7720-a334-4276-9531-a49fb9558275-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-t8psr\" (UID: \"017b7720-a334-4276-9531-a49fb9558275\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" Feb 25 13:43:08 crc kubenswrapper[4815]: I0225 13:43:08.077889 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/017b7720-a334-4276-9531-a49fb9558275-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-t8psr\" (UID: \"017b7720-a334-4276-9531-a49fb9558275\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" Feb 25 13:43:08 crc kubenswrapper[4815]: I0225 13:43:08.077936 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4frrm\" (UniqueName: \"kubernetes.io/projected/017b7720-a334-4276-9531-a49fb9558275-kube-api-access-4frrm\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-t8psr\" (UID: \"017b7720-a334-4276-9531-a49fb9558275\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" Feb 25 13:43:08 crc kubenswrapper[4815]: I0225 13:43:08.085363 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/017b7720-a334-4276-9531-a49fb9558275-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-t8psr\" (UID: \"017b7720-a334-4276-9531-a49fb9558275\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" Feb 25 13:43:08 crc kubenswrapper[4815]: I0225 13:43:08.095312 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/017b7720-a334-4276-9531-a49fb9558275-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-t8psr\" (UID: \"017b7720-a334-4276-9531-a49fb9558275\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" Feb 25 13:43:08 crc kubenswrapper[4815]: I0225 13:43:08.108375 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4frrm\" (UniqueName: \"kubernetes.io/projected/017b7720-a334-4276-9531-a49fb9558275-kube-api-access-4frrm\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-t8psr\" (UID: \"017b7720-a334-4276-9531-a49fb9558275\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" Feb 25 13:43:08 crc kubenswrapper[4815]: I0225 13:43:08.198704 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" Feb 25 13:43:08 crc kubenswrapper[4815]: I0225 13:43:08.711121 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr"] Feb 25 13:43:08 crc kubenswrapper[4815]: W0225 13:43:08.713835 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod017b7720_a334_4276_9531_a49fb9558275.slice/crio-76f747bbace77da909d3736defc291b3a841311252f68548efd6331ed722679b WatchSource:0}: Error finding container 76f747bbace77da909d3736defc291b3a841311252f68548efd6331ed722679b: Status 404 returned error can't find the container with id 76f747bbace77da909d3736defc291b3a841311252f68548efd6331ed722679b Feb 25 13:43:08 crc kubenswrapper[4815]: I0225 13:43:08.763162 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" event={"ID":"017b7720-a334-4276-9531-a49fb9558275","Type":"ContainerStarted","Data":"76f747bbace77da909d3736defc291b3a841311252f68548efd6331ed722679b"} Feb 25 13:43:09 crc kubenswrapper[4815]: I0225 13:43:09.773867 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" event={"ID":"017b7720-a334-4276-9531-a49fb9558275","Type":"ContainerStarted","Data":"78c9185139278142414a752ca85a2b2716584d367570ed56f81d604bfdf74078"} Feb 25 13:43:09 crc kubenswrapper[4815]: I0225 13:43:09.795346 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" podStartSLOduration=2.3901778240000002 podStartE2EDuration="2.794821065s" podCreationTimestamp="2026-02-25 13:43:07 +0000 UTC" firstStartedPulling="2026-02-25 13:43:08.717148733 +0000 UTC m=+1346.518246787" lastFinishedPulling="2026-02-25 13:43:09.121791984 +0000 UTC m=+1346.922890028" observedRunningTime="2026-02-25 13:43:09.789886405 +0000 UTC m=+1347.590984499" watchObservedRunningTime="2026-02-25 13:43:09.794821065 +0000 UTC m=+1347.595919119" Feb 25 13:43:11 crc kubenswrapper[4815]: I0225 13:43:11.796152 4815 generic.go:334] "Generic (PLEG): container finished" podID="017b7720-a334-4276-9531-a49fb9558275" containerID="78c9185139278142414a752ca85a2b2716584d367570ed56f81d604bfdf74078" exitCode=0 Feb 25 13:43:11 crc kubenswrapper[4815]: I0225 13:43:11.796187 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" event={"ID":"017b7720-a334-4276-9531-a49fb9558275","Type":"ContainerDied","Data":"78c9185139278142414a752ca85a2b2716584d367570ed56f81d604bfdf74078"} Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.510948 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.591776 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4frrm\" (UniqueName: \"kubernetes.io/projected/017b7720-a334-4276-9531-a49fb9558275-kube-api-access-4frrm\") pod \"017b7720-a334-4276-9531-a49fb9558275\" (UID: \"017b7720-a334-4276-9531-a49fb9558275\") " Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.591880 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/017b7720-a334-4276-9531-a49fb9558275-inventory\") pod \"017b7720-a334-4276-9531-a49fb9558275\" (UID: \"017b7720-a334-4276-9531-a49fb9558275\") " Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.591954 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/017b7720-a334-4276-9531-a49fb9558275-ssh-key-openstack-edpm-ipam\") pod \"017b7720-a334-4276-9531-a49fb9558275\" (UID: \"017b7720-a334-4276-9531-a49fb9558275\") " Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.612895 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/017b7720-a334-4276-9531-a49fb9558275-kube-api-access-4frrm" (OuterVolumeSpecName: "kube-api-access-4frrm") pod "017b7720-a334-4276-9531-a49fb9558275" (UID: "017b7720-a334-4276-9531-a49fb9558275"). InnerVolumeSpecName "kube-api-access-4frrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.637929 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/017b7720-a334-4276-9531-a49fb9558275-inventory" (OuterVolumeSpecName: "inventory") pod "017b7720-a334-4276-9531-a49fb9558275" (UID: "017b7720-a334-4276-9531-a49fb9558275"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.670659 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/017b7720-a334-4276-9531-a49fb9558275-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "017b7720-a334-4276-9531-a49fb9558275" (UID: "017b7720-a334-4276-9531-a49fb9558275"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.695715 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4frrm\" (UniqueName: \"kubernetes.io/projected/017b7720-a334-4276-9531-a49fb9558275-kube-api-access-4frrm\") on node \"crc\" DevicePath \"\"" Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.695751 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/017b7720-a334-4276-9531-a49fb9558275-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.695766 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/017b7720-a334-4276-9531-a49fb9558275-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.820535 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" event={"ID":"017b7720-a334-4276-9531-a49fb9558275","Type":"ContainerDied","Data":"76f747bbace77da909d3736defc291b3a841311252f68548efd6331ed722679b"} Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.820594 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76f747bbace77da909d3736defc291b3a841311252f68548efd6331ed722679b" Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.820757 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-t8psr" Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.906539 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz"] Feb 25 13:43:13 crc kubenswrapper[4815]: E0225 13:43:13.907145 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="017b7720-a334-4276-9531-a49fb9558275" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.907207 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="017b7720-a334-4276-9531-a49fb9558275" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.907413 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="017b7720-a334-4276-9531-a49fb9558275" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.908112 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.910346 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mmbj" Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.910385 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.911490 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.911723 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 25 13:43:13 crc kubenswrapper[4815]: I0225 13:43:13.936969 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz"] Feb 25 13:43:14 crc kubenswrapper[4815]: I0225 13:43:14.000986 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5fkw\" (UniqueName: \"kubernetes.io/projected/3a6985a9-1674-41ff-b441-23a83eff2148-kube-api-access-j5fkw\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz\" (UID: \"3a6985a9-1674-41ff-b441-23a83eff2148\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" Feb 25 13:43:14 crc kubenswrapper[4815]: I0225 13:43:14.001047 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz\" (UID: \"3a6985a9-1674-41ff-b441-23a83eff2148\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" Feb 25 13:43:14 crc kubenswrapper[4815]: I0225 13:43:14.001127 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz\" (UID: \"3a6985a9-1674-41ff-b441-23a83eff2148\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" Feb 25 13:43:14 crc kubenswrapper[4815]: I0225 13:43:14.001240 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz\" (UID: \"3a6985a9-1674-41ff-b441-23a83eff2148\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" Feb 25 13:43:14 crc kubenswrapper[4815]: I0225 13:43:14.103479 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5fkw\" (UniqueName: \"kubernetes.io/projected/3a6985a9-1674-41ff-b441-23a83eff2148-kube-api-access-j5fkw\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz\" (UID: \"3a6985a9-1674-41ff-b441-23a83eff2148\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" Feb 25 13:43:14 crc kubenswrapper[4815]: I0225 13:43:14.103575 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz\" (UID: \"3a6985a9-1674-41ff-b441-23a83eff2148\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" Feb 25 13:43:14 crc kubenswrapper[4815]: I0225 13:43:14.103614 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz\" (UID: \"3a6985a9-1674-41ff-b441-23a83eff2148\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" Feb 25 13:43:14 crc kubenswrapper[4815]: I0225 13:43:14.103690 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz\" (UID: \"3a6985a9-1674-41ff-b441-23a83eff2148\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" Feb 25 13:43:14 crc kubenswrapper[4815]: I0225 13:43:14.108212 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz\" (UID: \"3a6985a9-1674-41ff-b441-23a83eff2148\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" Feb 25 13:43:14 crc kubenswrapper[4815]: I0225 13:43:14.109128 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz\" (UID: \"3a6985a9-1674-41ff-b441-23a83eff2148\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" Feb 25 13:43:14 crc kubenswrapper[4815]: I0225 13:43:14.110694 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz\" (UID: \"3a6985a9-1674-41ff-b441-23a83eff2148\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" Feb 25 13:43:14 crc kubenswrapper[4815]: I0225 13:43:14.137313 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5fkw\" (UniqueName: \"kubernetes.io/projected/3a6985a9-1674-41ff-b441-23a83eff2148-kube-api-access-j5fkw\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz\" (UID: \"3a6985a9-1674-41ff-b441-23a83eff2148\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" Feb 25 13:43:14 crc kubenswrapper[4815]: I0225 13:43:14.225536 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" Feb 25 13:43:14 crc kubenswrapper[4815]: I0225 13:43:14.862646 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz"] Feb 25 13:43:15 crc kubenswrapper[4815]: I0225 13:43:15.846550 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" event={"ID":"3a6985a9-1674-41ff-b441-23a83eff2148","Type":"ContainerStarted","Data":"0436aaf0dbd5a6ae739f8097cf700105a672b8b981562794d8114146df3db7eb"} Feb 25 13:43:15 crc kubenswrapper[4815]: I0225 13:43:15.846892 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" event={"ID":"3a6985a9-1674-41ff-b441-23a83eff2148","Type":"ContainerStarted","Data":"77a68cee2e5c93077335c90691088506eecd6d98cc45531d78e6d3c272e5f36a"} Feb 25 13:43:15 crc kubenswrapper[4815]: I0225 13:43:15.879955 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" podStartSLOduration=2.463822478 podStartE2EDuration="2.879930667s" podCreationTimestamp="2026-02-25 13:43:13 +0000 UTC" firstStartedPulling="2026-02-25 13:43:14.866701867 +0000 UTC m=+1352.667799961" lastFinishedPulling="2026-02-25 13:43:15.282810096 +0000 UTC m=+1353.083908150" observedRunningTime="2026-02-25 13:43:15.868919881 +0000 UTC m=+1353.670017985" watchObservedRunningTime="2026-02-25 13:43:15.879930667 +0000 UTC m=+1353.681028731" Feb 25 13:43:21 crc kubenswrapper[4815]: I0225 13:43:21.342886 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:43:21 crc kubenswrapper[4815]: I0225 13:43:21.343483 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:43:21 crc kubenswrapper[4815]: I0225 13:43:21.343638 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:43:21 crc kubenswrapper[4815]: I0225 13:43:21.344658 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f21f73b0f5d16d6a34d568e8c21d1a82e30512997d6c1c659ca1e100017fbcdd"} pod="openshift-machine-config-operator/machine-config-daemon-frc75" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 13:43:21 crc kubenswrapper[4815]: I0225 13:43:21.344775 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" containerID="cri-o://f21f73b0f5d16d6a34d568e8c21d1a82e30512997d6c1c659ca1e100017fbcdd" gracePeriod=600 Feb 25 13:43:21 crc kubenswrapper[4815]: I0225 13:43:21.923030 4815 generic.go:334] "Generic (PLEG): container finished" podID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerID="f21f73b0f5d16d6a34d568e8c21d1a82e30512997d6c1c659ca1e100017fbcdd" exitCode=0 Feb 25 13:43:21 crc kubenswrapper[4815]: I0225 13:43:21.923099 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerDied","Data":"f21f73b0f5d16d6a34d568e8c21d1a82e30512997d6c1c659ca1e100017fbcdd"} Feb 25 13:43:21 crc kubenswrapper[4815]: I0225 13:43:21.923534 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerStarted","Data":"faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b"} Feb 25 13:43:21 crc kubenswrapper[4815]: I0225 13:43:21.923571 4815 scope.go:117] "RemoveContainer" containerID="90119f271263ed85c7a4bae34f95ffae1fb69e1997807ded5781ce1fee85b5cc" Feb 25 13:43:45 crc kubenswrapper[4815]: I0225 13:43:45.513107 4815 scope.go:117] "RemoveContainer" containerID="ebe6f94cd2653d2dbce26684653779141960ca6b8db847bde41f11f2653d6590" Feb 25 13:43:45 crc kubenswrapper[4815]: I0225 13:43:45.549067 4815 scope.go:117] "RemoveContainer" containerID="5d82924ff0f902481efdf03a6e242bf482ae47b4379d98e8299ecdfbb0c47469" Feb 25 13:43:45 crc kubenswrapper[4815]: I0225 13:43:45.627237 4815 scope.go:117] "RemoveContainer" containerID="9ce8c60b92b710cea3afd50b2e1693cb413c6423d29d6fa7e857c9247b4c3947" Feb 25 13:44:00 crc kubenswrapper[4815]: I0225 13:44:00.186184 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533784-xcml6"] Feb 25 13:44:00 crc kubenswrapper[4815]: I0225 13:44:00.189543 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533784-xcml6" Feb 25 13:44:00 crc kubenswrapper[4815]: I0225 13:44:00.192859 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:44:00 crc kubenswrapper[4815]: I0225 13:44:00.192859 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:44:00 crc kubenswrapper[4815]: I0225 13:44:00.193263 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:44:00 crc kubenswrapper[4815]: I0225 13:44:00.208668 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533784-xcml6"] Feb 25 13:44:00 crc kubenswrapper[4815]: I0225 13:44:00.246087 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp4jv\" (UniqueName: \"kubernetes.io/projected/ad0b5715-982d-4406-9815-d281ee2d5ebf-kube-api-access-zp4jv\") pod \"auto-csr-approver-29533784-xcml6\" (UID: \"ad0b5715-982d-4406-9815-d281ee2d5ebf\") " pod="openshift-infra/auto-csr-approver-29533784-xcml6" Feb 25 13:44:00 crc kubenswrapper[4815]: I0225 13:44:00.348277 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp4jv\" (UniqueName: \"kubernetes.io/projected/ad0b5715-982d-4406-9815-d281ee2d5ebf-kube-api-access-zp4jv\") pod \"auto-csr-approver-29533784-xcml6\" (UID: \"ad0b5715-982d-4406-9815-d281ee2d5ebf\") " pod="openshift-infra/auto-csr-approver-29533784-xcml6" Feb 25 13:44:00 crc kubenswrapper[4815]: I0225 13:44:00.371526 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp4jv\" (UniqueName: \"kubernetes.io/projected/ad0b5715-982d-4406-9815-d281ee2d5ebf-kube-api-access-zp4jv\") pod \"auto-csr-approver-29533784-xcml6\" (UID: \"ad0b5715-982d-4406-9815-d281ee2d5ebf\") " pod="openshift-infra/auto-csr-approver-29533784-xcml6" Feb 25 13:44:00 crc kubenswrapper[4815]: I0225 13:44:00.511588 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533784-xcml6" Feb 25 13:44:00 crc kubenswrapper[4815]: I0225 13:44:00.972855 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533784-xcml6"] Feb 25 13:44:00 crc kubenswrapper[4815]: I0225 13:44:00.978143 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 13:44:01 crc kubenswrapper[4815]: I0225 13:44:01.335106 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533784-xcml6" event={"ID":"ad0b5715-982d-4406-9815-d281ee2d5ebf","Type":"ContainerStarted","Data":"3a085e04078a9050113b051a1656334967d2a22937635243f603765cbc03d54e"} Feb 25 13:44:02 crc kubenswrapper[4815]: I0225 13:44:02.346059 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533784-xcml6" event={"ID":"ad0b5715-982d-4406-9815-d281ee2d5ebf","Type":"ContainerStarted","Data":"c9a93486d2ab0ff1a07236d6049b2378f1b98706e40d3b5bdace6972d3e041dc"} Feb 25 13:44:02 crc kubenswrapper[4815]: I0225 13:44:02.362959 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533784-xcml6" podStartSLOduration=1.4022375870000001 podStartE2EDuration="2.362938967s" podCreationTimestamp="2026-02-25 13:44:00 +0000 UTC" firstStartedPulling="2026-02-25 13:44:00.977901972 +0000 UTC m=+1398.779000026" lastFinishedPulling="2026-02-25 13:44:01.938603302 +0000 UTC m=+1399.739701406" observedRunningTime="2026-02-25 13:44:02.361675358 +0000 UTC m=+1400.162773412" watchObservedRunningTime="2026-02-25 13:44:02.362938967 +0000 UTC m=+1400.164037021" Feb 25 13:44:03 crc kubenswrapper[4815]: I0225 13:44:03.362298 4815 generic.go:334] "Generic (PLEG): container finished" podID="ad0b5715-982d-4406-9815-d281ee2d5ebf" containerID="c9a93486d2ab0ff1a07236d6049b2378f1b98706e40d3b5bdace6972d3e041dc" exitCode=0 Feb 25 13:44:03 crc kubenswrapper[4815]: I0225 13:44:03.362428 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533784-xcml6" event={"ID":"ad0b5715-982d-4406-9815-d281ee2d5ebf","Type":"ContainerDied","Data":"c9a93486d2ab0ff1a07236d6049b2378f1b98706e40d3b5bdace6972d3e041dc"} Feb 25 13:44:04 crc kubenswrapper[4815]: I0225 13:44:04.714165 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533784-xcml6" Feb 25 13:44:04 crc kubenswrapper[4815]: I0225 13:44:04.738381 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp4jv\" (UniqueName: \"kubernetes.io/projected/ad0b5715-982d-4406-9815-d281ee2d5ebf-kube-api-access-zp4jv\") pod \"ad0b5715-982d-4406-9815-d281ee2d5ebf\" (UID: \"ad0b5715-982d-4406-9815-d281ee2d5ebf\") " Feb 25 13:44:04 crc kubenswrapper[4815]: I0225 13:44:04.744018 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad0b5715-982d-4406-9815-d281ee2d5ebf-kube-api-access-zp4jv" (OuterVolumeSpecName: "kube-api-access-zp4jv") pod "ad0b5715-982d-4406-9815-d281ee2d5ebf" (UID: "ad0b5715-982d-4406-9815-d281ee2d5ebf"). InnerVolumeSpecName "kube-api-access-zp4jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:44:04 crc kubenswrapper[4815]: I0225 13:44:04.841211 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp4jv\" (UniqueName: \"kubernetes.io/projected/ad0b5715-982d-4406-9815-d281ee2d5ebf-kube-api-access-zp4jv\") on node \"crc\" DevicePath \"\"" Feb 25 13:44:05 crc kubenswrapper[4815]: I0225 13:44:05.390277 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533784-xcml6" event={"ID":"ad0b5715-982d-4406-9815-d281ee2d5ebf","Type":"ContainerDied","Data":"3a085e04078a9050113b051a1656334967d2a22937635243f603765cbc03d54e"} Feb 25 13:44:05 crc kubenswrapper[4815]: I0225 13:44:05.390319 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a085e04078a9050113b051a1656334967d2a22937635243f603765cbc03d54e" Feb 25 13:44:05 crc kubenswrapper[4815]: I0225 13:44:05.390368 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533784-xcml6" Feb 25 13:44:05 crc kubenswrapper[4815]: I0225 13:44:05.472147 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533778-xbfls"] Feb 25 13:44:05 crc kubenswrapper[4815]: I0225 13:44:05.479807 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533778-xbfls"] Feb 25 13:44:06 crc kubenswrapper[4815]: I0225 13:44:06.948333 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5139ef47-6709-46f0-8f06-46cadbbb36b1" path="/var/lib/kubelet/pods/5139ef47-6709-46f0-8f06-46cadbbb36b1/volumes" Feb 25 13:44:45 crc kubenswrapper[4815]: I0225 13:44:45.730339 4815 scope.go:117] "RemoveContainer" containerID="d09a900a7289126fcf20dcc95e26d94934df340a2b7e2ea2c89358cd26e4b697" Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.152481 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z"] Feb 25 13:45:00 crc kubenswrapper[4815]: E0225 13:45:00.153971 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad0b5715-982d-4406-9815-d281ee2d5ebf" containerName="oc" Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.153991 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad0b5715-982d-4406-9815-d281ee2d5ebf" containerName="oc" Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.154239 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad0b5715-982d-4406-9815-d281ee2d5ebf" containerName="oc" Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.155015 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z" Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.157854 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.158168 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.175013 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z"] Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.346954 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbsdz\" (UniqueName: \"kubernetes.io/projected/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-kube-api-access-nbsdz\") pod \"collect-profiles-29533785-gtq8z\" (UID: \"84bf216f-9085-4cfe-a8e1-986d75f0ba5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z" Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.347077 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-secret-volume\") pod \"collect-profiles-29533785-gtq8z\" (UID: \"84bf216f-9085-4cfe-a8e1-986d75f0ba5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z" Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.347172 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-config-volume\") pod \"collect-profiles-29533785-gtq8z\" (UID: \"84bf216f-9085-4cfe-a8e1-986d75f0ba5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z" Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.448751 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-config-volume\") pod \"collect-profiles-29533785-gtq8z\" (UID: \"84bf216f-9085-4cfe-a8e1-986d75f0ba5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z" Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.448829 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbsdz\" (UniqueName: \"kubernetes.io/projected/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-kube-api-access-nbsdz\") pod \"collect-profiles-29533785-gtq8z\" (UID: \"84bf216f-9085-4cfe-a8e1-986d75f0ba5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z" Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.448955 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-secret-volume\") pod \"collect-profiles-29533785-gtq8z\" (UID: \"84bf216f-9085-4cfe-a8e1-986d75f0ba5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z" Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.449681 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-config-volume\") pod \"collect-profiles-29533785-gtq8z\" (UID: \"84bf216f-9085-4cfe-a8e1-986d75f0ba5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z" Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.455347 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-secret-volume\") pod \"collect-profiles-29533785-gtq8z\" (UID: \"84bf216f-9085-4cfe-a8e1-986d75f0ba5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z" Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.469756 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbsdz\" (UniqueName: \"kubernetes.io/projected/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-kube-api-access-nbsdz\") pod \"collect-profiles-29533785-gtq8z\" (UID: \"84bf216f-9085-4cfe-a8e1-986d75f0ba5d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z" Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.483718 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z" Feb 25 13:45:00 crc kubenswrapper[4815]: I0225 13:45:00.980297 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z"] Feb 25 13:45:01 crc kubenswrapper[4815]: I0225 13:45:01.007995 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z" event={"ID":"84bf216f-9085-4cfe-a8e1-986d75f0ba5d","Type":"ContainerStarted","Data":"543e511797405574066a20d3eff933f57f239c88a9fdb1cbfabb48fb904e9683"} Feb 25 13:45:02 crc kubenswrapper[4815]: I0225 13:45:02.020751 4815 generic.go:334] "Generic (PLEG): container finished" podID="84bf216f-9085-4cfe-a8e1-986d75f0ba5d" containerID="5e1bfe13f9a5a0ee6871e4286e67509181be30bbf5c498a31397193ccd5a5c0d" exitCode=0 Feb 25 13:45:02 crc kubenswrapper[4815]: I0225 13:45:02.020807 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z" event={"ID":"84bf216f-9085-4cfe-a8e1-986d75f0ba5d","Type":"ContainerDied","Data":"5e1bfe13f9a5a0ee6871e4286e67509181be30bbf5c498a31397193ccd5a5c0d"} Feb 25 13:45:03 crc kubenswrapper[4815]: I0225 13:45:03.404673 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z" Feb 25 13:45:03 crc kubenswrapper[4815]: I0225 13:45:03.505302 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-config-volume\") pod \"84bf216f-9085-4cfe-a8e1-986d75f0ba5d\" (UID: \"84bf216f-9085-4cfe-a8e1-986d75f0ba5d\") " Feb 25 13:45:03 crc kubenswrapper[4815]: I0225 13:45:03.505348 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbsdz\" (UniqueName: \"kubernetes.io/projected/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-kube-api-access-nbsdz\") pod \"84bf216f-9085-4cfe-a8e1-986d75f0ba5d\" (UID: \"84bf216f-9085-4cfe-a8e1-986d75f0ba5d\") " Feb 25 13:45:03 crc kubenswrapper[4815]: I0225 13:45:03.505372 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-secret-volume\") pod \"84bf216f-9085-4cfe-a8e1-986d75f0ba5d\" (UID: \"84bf216f-9085-4cfe-a8e1-986d75f0ba5d\") " Feb 25 13:45:03 crc kubenswrapper[4815]: I0225 13:45:03.505899 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-config-volume" (OuterVolumeSpecName: "config-volume") pod "84bf216f-9085-4cfe-a8e1-986d75f0ba5d" (UID: "84bf216f-9085-4cfe-a8e1-986d75f0ba5d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:45:03 crc kubenswrapper[4815]: I0225 13:45:03.506269 4815 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 13:45:03 crc kubenswrapper[4815]: I0225 13:45:03.512235 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-kube-api-access-nbsdz" (OuterVolumeSpecName: "kube-api-access-nbsdz") pod "84bf216f-9085-4cfe-a8e1-986d75f0ba5d" (UID: "84bf216f-9085-4cfe-a8e1-986d75f0ba5d"). InnerVolumeSpecName "kube-api-access-nbsdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:45:03 crc kubenswrapper[4815]: I0225 13:45:03.513701 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "84bf216f-9085-4cfe-a8e1-986d75f0ba5d" (UID: "84bf216f-9085-4cfe-a8e1-986d75f0ba5d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:45:03 crc kubenswrapper[4815]: I0225 13:45:03.610151 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbsdz\" (UniqueName: \"kubernetes.io/projected/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-kube-api-access-nbsdz\") on node \"crc\" DevicePath \"\"" Feb 25 13:45:03 crc kubenswrapper[4815]: I0225 13:45:03.610478 4815 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/84bf216f-9085-4cfe-a8e1-986d75f0ba5d-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.041956 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z" event={"ID":"84bf216f-9085-4cfe-a8e1-986d75f0ba5d","Type":"ContainerDied","Data":"543e511797405574066a20d3eff933f57f239c88a9fdb1cbfabb48fb904e9683"} Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.042012 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="543e511797405574066a20d3eff933f57f239c88a9fdb1cbfabb48fb904e9683" Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.042080 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533785-gtq8z" Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.447187 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2wt6f"] Feb 25 13:45:04 crc kubenswrapper[4815]: E0225 13:45:04.447788 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84bf216f-9085-4cfe-a8e1-986d75f0ba5d" containerName="collect-profiles" Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.447804 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="84bf216f-9085-4cfe-a8e1-986d75f0ba5d" containerName="collect-profiles" Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.448064 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="84bf216f-9085-4cfe-a8e1-986d75f0ba5d" containerName="collect-profiles" Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.449651 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2wt6f" Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.470069 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2wt6f"] Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.627981 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b842e52e-04d7-43b5-99e2-2816c1d12eed-utilities\") pod \"redhat-marketplace-2wt6f\" (UID: \"b842e52e-04d7-43b5-99e2-2816c1d12eed\") " pod="openshift-marketplace/redhat-marketplace-2wt6f" Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.628070 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8rwx\" (UniqueName: \"kubernetes.io/projected/b842e52e-04d7-43b5-99e2-2816c1d12eed-kube-api-access-z8rwx\") pod \"redhat-marketplace-2wt6f\" (UID: \"b842e52e-04d7-43b5-99e2-2816c1d12eed\") " pod="openshift-marketplace/redhat-marketplace-2wt6f" Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.628227 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b842e52e-04d7-43b5-99e2-2816c1d12eed-catalog-content\") pod \"redhat-marketplace-2wt6f\" (UID: \"b842e52e-04d7-43b5-99e2-2816c1d12eed\") " pod="openshift-marketplace/redhat-marketplace-2wt6f" Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.730258 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b842e52e-04d7-43b5-99e2-2816c1d12eed-catalog-content\") pod \"redhat-marketplace-2wt6f\" (UID: \"b842e52e-04d7-43b5-99e2-2816c1d12eed\") " pod="openshift-marketplace/redhat-marketplace-2wt6f" Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.730380 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b842e52e-04d7-43b5-99e2-2816c1d12eed-utilities\") pod \"redhat-marketplace-2wt6f\" (UID: \"b842e52e-04d7-43b5-99e2-2816c1d12eed\") " pod="openshift-marketplace/redhat-marketplace-2wt6f" Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.730459 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8rwx\" (UniqueName: \"kubernetes.io/projected/b842e52e-04d7-43b5-99e2-2816c1d12eed-kube-api-access-z8rwx\") pod \"redhat-marketplace-2wt6f\" (UID: \"b842e52e-04d7-43b5-99e2-2816c1d12eed\") " pod="openshift-marketplace/redhat-marketplace-2wt6f" Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.730761 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b842e52e-04d7-43b5-99e2-2816c1d12eed-catalog-content\") pod \"redhat-marketplace-2wt6f\" (UID: \"b842e52e-04d7-43b5-99e2-2816c1d12eed\") " pod="openshift-marketplace/redhat-marketplace-2wt6f" Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.730832 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b842e52e-04d7-43b5-99e2-2816c1d12eed-utilities\") pod \"redhat-marketplace-2wt6f\" (UID: \"b842e52e-04d7-43b5-99e2-2816c1d12eed\") " pod="openshift-marketplace/redhat-marketplace-2wt6f" Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.750669 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8rwx\" (UniqueName: \"kubernetes.io/projected/b842e52e-04d7-43b5-99e2-2816c1d12eed-kube-api-access-z8rwx\") pod \"redhat-marketplace-2wt6f\" (UID: \"b842e52e-04d7-43b5-99e2-2816c1d12eed\") " pod="openshift-marketplace/redhat-marketplace-2wt6f" Feb 25 13:45:04 crc kubenswrapper[4815]: I0225 13:45:04.769589 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2wt6f" Feb 25 13:45:05 crc kubenswrapper[4815]: I0225 13:45:05.273110 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2wt6f"] Feb 25 13:45:06 crc kubenswrapper[4815]: I0225 13:45:06.067097 4815 generic.go:334] "Generic (PLEG): container finished" podID="b842e52e-04d7-43b5-99e2-2816c1d12eed" containerID="d1c5ba34bbd46d598ae2967226f7e3fee21aa119443d106843fc9a1f95ff34c6" exitCode=0 Feb 25 13:45:06 crc kubenswrapper[4815]: I0225 13:45:06.067254 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2wt6f" event={"ID":"b842e52e-04d7-43b5-99e2-2816c1d12eed","Type":"ContainerDied","Data":"d1c5ba34bbd46d598ae2967226f7e3fee21aa119443d106843fc9a1f95ff34c6"} Feb 25 13:45:06 crc kubenswrapper[4815]: I0225 13:45:06.067939 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2wt6f" event={"ID":"b842e52e-04d7-43b5-99e2-2816c1d12eed","Type":"ContainerStarted","Data":"01208ea6ac2b934d9ac1186fe87a4ec5c3e70c80b3e17c94c040292ad0095486"} Feb 25 13:45:08 crc kubenswrapper[4815]: I0225 13:45:08.118420 4815 generic.go:334] "Generic (PLEG): container finished" podID="b842e52e-04d7-43b5-99e2-2816c1d12eed" containerID="aff711d7a2f6a4e830eab771ab62c1c164e2284962cb8971794129bb4fadbf73" exitCode=0 Feb 25 13:45:08 crc kubenswrapper[4815]: I0225 13:45:08.118554 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2wt6f" event={"ID":"b842e52e-04d7-43b5-99e2-2816c1d12eed","Type":"ContainerDied","Data":"aff711d7a2f6a4e830eab771ab62c1c164e2284962cb8971794129bb4fadbf73"} Feb 25 13:45:09 crc kubenswrapper[4815]: I0225 13:45:09.128477 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2wt6f" event={"ID":"b842e52e-04d7-43b5-99e2-2816c1d12eed","Type":"ContainerStarted","Data":"5c03142a69353dceb6b8366ed0a25b69dacfda49f14a188c9e581802bb0cb793"} Feb 25 13:45:09 crc kubenswrapper[4815]: I0225 13:45:09.146547 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2wt6f" podStartSLOduration=2.713806543 podStartE2EDuration="5.146501459s" podCreationTimestamp="2026-02-25 13:45:04 +0000 UTC" firstStartedPulling="2026-02-25 13:45:06.069502571 +0000 UTC m=+1463.870600625" lastFinishedPulling="2026-02-25 13:45:08.502197477 +0000 UTC m=+1466.303295541" observedRunningTime="2026-02-25 13:45:09.142934768 +0000 UTC m=+1466.944032832" watchObservedRunningTime="2026-02-25 13:45:09.146501459 +0000 UTC m=+1466.947599513" Feb 25 13:45:14 crc kubenswrapper[4815]: I0225 13:45:14.770582 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2wt6f" Feb 25 13:45:14 crc kubenswrapper[4815]: I0225 13:45:14.771362 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2wt6f" Feb 25 13:45:14 crc kubenswrapper[4815]: I0225 13:45:14.844608 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2wt6f" Feb 25 13:45:15 crc kubenswrapper[4815]: I0225 13:45:15.256710 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2wt6f" Feb 25 13:45:15 crc kubenswrapper[4815]: I0225 13:45:15.309876 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2wt6f"] Feb 25 13:45:17 crc kubenswrapper[4815]: I0225 13:45:17.229927 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2wt6f" podUID="b842e52e-04d7-43b5-99e2-2816c1d12eed" containerName="registry-server" containerID="cri-o://5c03142a69353dceb6b8366ed0a25b69dacfda49f14a188c9e581802bb0cb793" gracePeriod=2 Feb 25 13:45:17 crc kubenswrapper[4815]: I0225 13:45:17.803918 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2wt6f" Feb 25 13:45:17 crc kubenswrapper[4815]: I0225 13:45:17.898326 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b842e52e-04d7-43b5-99e2-2816c1d12eed-utilities\") pod \"b842e52e-04d7-43b5-99e2-2816c1d12eed\" (UID: \"b842e52e-04d7-43b5-99e2-2816c1d12eed\") " Feb 25 13:45:17 crc kubenswrapper[4815]: I0225 13:45:17.898685 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b842e52e-04d7-43b5-99e2-2816c1d12eed-catalog-content\") pod \"b842e52e-04d7-43b5-99e2-2816c1d12eed\" (UID: \"b842e52e-04d7-43b5-99e2-2816c1d12eed\") " Feb 25 13:45:17 crc kubenswrapper[4815]: I0225 13:45:17.898950 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8rwx\" (UniqueName: \"kubernetes.io/projected/b842e52e-04d7-43b5-99e2-2816c1d12eed-kube-api-access-z8rwx\") pod \"b842e52e-04d7-43b5-99e2-2816c1d12eed\" (UID: \"b842e52e-04d7-43b5-99e2-2816c1d12eed\") " Feb 25 13:45:17 crc kubenswrapper[4815]: I0225 13:45:17.899541 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b842e52e-04d7-43b5-99e2-2816c1d12eed-utilities" (OuterVolumeSpecName: "utilities") pod "b842e52e-04d7-43b5-99e2-2816c1d12eed" (UID: "b842e52e-04d7-43b5-99e2-2816c1d12eed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:45:17 crc kubenswrapper[4815]: I0225 13:45:17.912697 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b842e52e-04d7-43b5-99e2-2816c1d12eed-kube-api-access-z8rwx" (OuterVolumeSpecName: "kube-api-access-z8rwx") pod "b842e52e-04d7-43b5-99e2-2816c1d12eed" (UID: "b842e52e-04d7-43b5-99e2-2816c1d12eed"). InnerVolumeSpecName "kube-api-access-z8rwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:45:17 crc kubenswrapper[4815]: I0225 13:45:17.934057 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b842e52e-04d7-43b5-99e2-2816c1d12eed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b842e52e-04d7-43b5-99e2-2816c1d12eed" (UID: "b842e52e-04d7-43b5-99e2-2816c1d12eed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.002166 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8rwx\" (UniqueName: \"kubernetes.io/projected/b842e52e-04d7-43b5-99e2-2816c1d12eed-kube-api-access-z8rwx\") on node \"crc\" DevicePath \"\"" Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.002204 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b842e52e-04d7-43b5-99e2-2816c1d12eed-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.002219 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b842e52e-04d7-43b5-99e2-2816c1d12eed-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.246458 4815 generic.go:334] "Generic (PLEG): container finished" podID="b842e52e-04d7-43b5-99e2-2816c1d12eed" containerID="5c03142a69353dceb6b8366ed0a25b69dacfda49f14a188c9e581802bb0cb793" exitCode=0 Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.246547 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2wt6f" event={"ID":"b842e52e-04d7-43b5-99e2-2816c1d12eed","Type":"ContainerDied","Data":"5c03142a69353dceb6b8366ed0a25b69dacfda49f14a188c9e581802bb0cb793"} Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.246627 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2wt6f" event={"ID":"b842e52e-04d7-43b5-99e2-2816c1d12eed","Type":"ContainerDied","Data":"01208ea6ac2b934d9ac1186fe87a4ec5c3e70c80b3e17c94c040292ad0095486"} Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.246649 4815 scope.go:117] "RemoveContainer" containerID="5c03142a69353dceb6b8366ed0a25b69dacfda49f14a188c9e581802bb0cb793" Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.246574 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2wt6f" Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.273758 4815 scope.go:117] "RemoveContainer" containerID="aff711d7a2f6a4e830eab771ab62c1c164e2284962cb8971794129bb4fadbf73" Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.294406 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2wt6f"] Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.309140 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2wt6f"] Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.317710 4815 scope.go:117] "RemoveContainer" containerID="d1c5ba34bbd46d598ae2967226f7e3fee21aa119443d106843fc9a1f95ff34c6" Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.362249 4815 scope.go:117] "RemoveContainer" containerID="5c03142a69353dceb6b8366ed0a25b69dacfda49f14a188c9e581802bb0cb793" Feb 25 13:45:18 crc kubenswrapper[4815]: E0225 13:45:18.362735 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c03142a69353dceb6b8366ed0a25b69dacfda49f14a188c9e581802bb0cb793\": container with ID starting with 5c03142a69353dceb6b8366ed0a25b69dacfda49f14a188c9e581802bb0cb793 not found: ID does not exist" containerID="5c03142a69353dceb6b8366ed0a25b69dacfda49f14a188c9e581802bb0cb793" Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.362762 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c03142a69353dceb6b8366ed0a25b69dacfda49f14a188c9e581802bb0cb793"} err="failed to get container status \"5c03142a69353dceb6b8366ed0a25b69dacfda49f14a188c9e581802bb0cb793\": rpc error: code = NotFound desc = could not find container \"5c03142a69353dceb6b8366ed0a25b69dacfda49f14a188c9e581802bb0cb793\": container with ID starting with 5c03142a69353dceb6b8366ed0a25b69dacfda49f14a188c9e581802bb0cb793 not found: ID does not exist" Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.362782 4815 scope.go:117] "RemoveContainer" containerID="aff711d7a2f6a4e830eab771ab62c1c164e2284962cb8971794129bb4fadbf73" Feb 25 13:45:18 crc kubenswrapper[4815]: E0225 13:45:18.363250 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aff711d7a2f6a4e830eab771ab62c1c164e2284962cb8971794129bb4fadbf73\": container with ID starting with aff711d7a2f6a4e830eab771ab62c1c164e2284962cb8971794129bb4fadbf73 not found: ID does not exist" containerID="aff711d7a2f6a4e830eab771ab62c1c164e2284962cb8971794129bb4fadbf73" Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.363356 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aff711d7a2f6a4e830eab771ab62c1c164e2284962cb8971794129bb4fadbf73"} err="failed to get container status \"aff711d7a2f6a4e830eab771ab62c1c164e2284962cb8971794129bb4fadbf73\": rpc error: code = NotFound desc = could not find container \"aff711d7a2f6a4e830eab771ab62c1c164e2284962cb8971794129bb4fadbf73\": container with ID starting with aff711d7a2f6a4e830eab771ab62c1c164e2284962cb8971794129bb4fadbf73 not found: ID does not exist" Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.363425 4815 scope.go:117] "RemoveContainer" containerID="d1c5ba34bbd46d598ae2967226f7e3fee21aa119443d106843fc9a1f95ff34c6" Feb 25 13:45:18 crc kubenswrapper[4815]: E0225 13:45:18.364101 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1c5ba34bbd46d598ae2967226f7e3fee21aa119443d106843fc9a1f95ff34c6\": container with ID starting with d1c5ba34bbd46d598ae2967226f7e3fee21aa119443d106843fc9a1f95ff34c6 not found: ID does not exist" containerID="d1c5ba34bbd46d598ae2967226f7e3fee21aa119443d106843fc9a1f95ff34c6" Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.364206 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1c5ba34bbd46d598ae2967226f7e3fee21aa119443d106843fc9a1f95ff34c6"} err="failed to get container status \"d1c5ba34bbd46d598ae2967226f7e3fee21aa119443d106843fc9a1f95ff34c6\": rpc error: code = NotFound desc = could not find container \"d1c5ba34bbd46d598ae2967226f7e3fee21aa119443d106843fc9a1f95ff34c6\": container with ID starting with d1c5ba34bbd46d598ae2967226f7e3fee21aa119443d106843fc9a1f95ff34c6 not found: ID does not exist" Feb 25 13:45:18 crc kubenswrapper[4815]: I0225 13:45:18.948125 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b842e52e-04d7-43b5-99e2-2816c1d12eed" path="/var/lib/kubelet/pods/b842e52e-04d7-43b5-99e2-2816c1d12eed/volumes" Feb 25 13:45:21 crc kubenswrapper[4815]: I0225 13:45:21.342040 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:45:21 crc kubenswrapper[4815]: I0225 13:45:21.342450 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:45:29 crc kubenswrapper[4815]: I0225 13:45:29.547829 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z6fgs"] Feb 25 13:45:29 crc kubenswrapper[4815]: E0225 13:45:29.551630 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b842e52e-04d7-43b5-99e2-2816c1d12eed" containerName="extract-content" Feb 25 13:45:29 crc kubenswrapper[4815]: I0225 13:45:29.551664 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b842e52e-04d7-43b5-99e2-2816c1d12eed" containerName="extract-content" Feb 25 13:45:29 crc kubenswrapper[4815]: E0225 13:45:29.551679 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b842e52e-04d7-43b5-99e2-2816c1d12eed" containerName="registry-server" Feb 25 13:45:29 crc kubenswrapper[4815]: I0225 13:45:29.551687 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b842e52e-04d7-43b5-99e2-2816c1d12eed" containerName="registry-server" Feb 25 13:45:29 crc kubenswrapper[4815]: E0225 13:45:29.551704 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b842e52e-04d7-43b5-99e2-2816c1d12eed" containerName="extract-utilities" Feb 25 13:45:29 crc kubenswrapper[4815]: I0225 13:45:29.551711 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b842e52e-04d7-43b5-99e2-2816c1d12eed" containerName="extract-utilities" Feb 25 13:45:29 crc kubenswrapper[4815]: I0225 13:45:29.552160 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b842e52e-04d7-43b5-99e2-2816c1d12eed" containerName="registry-server" Feb 25 13:45:29 crc kubenswrapper[4815]: I0225 13:45:29.555200 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z6fgs" Feb 25 13:45:29 crc kubenswrapper[4815]: I0225 13:45:29.589607 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z6fgs"] Feb 25 13:45:29 crc kubenswrapper[4815]: I0225 13:45:29.652526 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/347a4d3a-d914-42b3-9239-ed7526039de6-catalog-content\") pod \"community-operators-z6fgs\" (UID: \"347a4d3a-d914-42b3-9239-ed7526039de6\") " pod="openshift-marketplace/community-operators-z6fgs" Feb 25 13:45:29 crc kubenswrapper[4815]: I0225 13:45:29.652627 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/347a4d3a-d914-42b3-9239-ed7526039de6-utilities\") pod \"community-operators-z6fgs\" (UID: \"347a4d3a-d914-42b3-9239-ed7526039de6\") " pod="openshift-marketplace/community-operators-z6fgs" Feb 25 13:45:29 crc kubenswrapper[4815]: I0225 13:45:29.652679 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fj52\" (UniqueName: \"kubernetes.io/projected/347a4d3a-d914-42b3-9239-ed7526039de6-kube-api-access-5fj52\") pod \"community-operators-z6fgs\" (UID: \"347a4d3a-d914-42b3-9239-ed7526039de6\") " pod="openshift-marketplace/community-operators-z6fgs" Feb 25 13:45:29 crc kubenswrapper[4815]: I0225 13:45:29.754653 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/347a4d3a-d914-42b3-9239-ed7526039de6-catalog-content\") pod \"community-operators-z6fgs\" (UID: \"347a4d3a-d914-42b3-9239-ed7526039de6\") " pod="openshift-marketplace/community-operators-z6fgs" Feb 25 13:45:29 crc kubenswrapper[4815]: I0225 13:45:29.754737 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/347a4d3a-d914-42b3-9239-ed7526039de6-utilities\") pod \"community-operators-z6fgs\" (UID: \"347a4d3a-d914-42b3-9239-ed7526039de6\") " pod="openshift-marketplace/community-operators-z6fgs" Feb 25 13:45:29 crc kubenswrapper[4815]: I0225 13:45:29.754774 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fj52\" (UniqueName: \"kubernetes.io/projected/347a4d3a-d914-42b3-9239-ed7526039de6-kube-api-access-5fj52\") pod \"community-operators-z6fgs\" (UID: \"347a4d3a-d914-42b3-9239-ed7526039de6\") " pod="openshift-marketplace/community-operators-z6fgs" Feb 25 13:45:29 crc kubenswrapper[4815]: I0225 13:45:29.755327 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/347a4d3a-d914-42b3-9239-ed7526039de6-catalog-content\") pod \"community-operators-z6fgs\" (UID: \"347a4d3a-d914-42b3-9239-ed7526039de6\") " pod="openshift-marketplace/community-operators-z6fgs" Feb 25 13:45:29 crc kubenswrapper[4815]: I0225 13:45:29.755342 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/347a4d3a-d914-42b3-9239-ed7526039de6-utilities\") pod \"community-operators-z6fgs\" (UID: \"347a4d3a-d914-42b3-9239-ed7526039de6\") " pod="openshift-marketplace/community-operators-z6fgs" Feb 25 13:45:29 crc kubenswrapper[4815]: I0225 13:45:29.781360 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fj52\" (UniqueName: \"kubernetes.io/projected/347a4d3a-d914-42b3-9239-ed7526039de6-kube-api-access-5fj52\") pod \"community-operators-z6fgs\" (UID: \"347a4d3a-d914-42b3-9239-ed7526039de6\") " pod="openshift-marketplace/community-operators-z6fgs" Feb 25 13:45:29 crc kubenswrapper[4815]: I0225 13:45:29.895414 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z6fgs" Feb 25 13:45:30 crc kubenswrapper[4815]: I0225 13:45:30.435819 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z6fgs"] Feb 25 13:45:31 crc kubenswrapper[4815]: I0225 13:45:31.412694 4815 generic.go:334] "Generic (PLEG): container finished" podID="347a4d3a-d914-42b3-9239-ed7526039de6" containerID="7fc8ac75f262c49f18ee595b211b26795dd9a2dd0f66ebefb46a887bddb9615a" exitCode=0 Feb 25 13:45:31 crc kubenswrapper[4815]: I0225 13:45:31.412772 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6fgs" event={"ID":"347a4d3a-d914-42b3-9239-ed7526039de6","Type":"ContainerDied","Data":"7fc8ac75f262c49f18ee595b211b26795dd9a2dd0f66ebefb46a887bddb9615a"} Feb 25 13:45:31 crc kubenswrapper[4815]: I0225 13:45:31.412946 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6fgs" event={"ID":"347a4d3a-d914-42b3-9239-ed7526039de6","Type":"ContainerStarted","Data":"c9bc5df069a4771edb7477ada74470555936d6f3cb5b35ee9068874c5af57ece"} Feb 25 13:45:35 crc kubenswrapper[4815]: I0225 13:45:35.460407 4815 generic.go:334] "Generic (PLEG): container finished" podID="347a4d3a-d914-42b3-9239-ed7526039de6" containerID="dd5b246b047d8e169f03ee3c49d659feaa1ac1c8604edc594d0330839a9e2d39" exitCode=0 Feb 25 13:45:35 crc kubenswrapper[4815]: I0225 13:45:35.461233 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6fgs" event={"ID":"347a4d3a-d914-42b3-9239-ed7526039de6","Type":"ContainerDied","Data":"dd5b246b047d8e169f03ee3c49d659feaa1ac1c8604edc594d0330839a9e2d39"} Feb 25 13:45:37 crc kubenswrapper[4815]: I0225 13:45:37.355533 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6rt2d"] Feb 25 13:45:37 crc kubenswrapper[4815]: I0225 13:45:37.358102 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6rt2d" Feb 25 13:45:37 crc kubenswrapper[4815]: I0225 13:45:37.421581 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6rt2d"] Feb 25 13:45:37 crc kubenswrapper[4815]: I0225 13:45:37.541646 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z6fgs" event={"ID":"347a4d3a-d914-42b3-9239-ed7526039de6","Type":"ContainerStarted","Data":"1d3b9cf1edf18684164ad0b975722bdc98e31552186a2fb548fdd1c6dace9b92"} Feb 25 13:45:37 crc kubenswrapper[4815]: I0225 13:45:37.546375 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a84a587-7d0c-4201-b618-2e2caa068fbb-utilities\") pod \"redhat-operators-6rt2d\" (UID: \"2a84a587-7d0c-4201-b618-2e2caa068fbb\") " pod="openshift-marketplace/redhat-operators-6rt2d" Feb 25 13:45:37 crc kubenswrapper[4815]: I0225 13:45:37.546430 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l54g8\" (UniqueName: \"kubernetes.io/projected/2a84a587-7d0c-4201-b618-2e2caa068fbb-kube-api-access-l54g8\") pod \"redhat-operators-6rt2d\" (UID: \"2a84a587-7d0c-4201-b618-2e2caa068fbb\") " pod="openshift-marketplace/redhat-operators-6rt2d" Feb 25 13:45:37 crc kubenswrapper[4815]: I0225 13:45:37.546453 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a84a587-7d0c-4201-b618-2e2caa068fbb-catalog-content\") pod \"redhat-operators-6rt2d\" (UID: \"2a84a587-7d0c-4201-b618-2e2caa068fbb\") " pod="openshift-marketplace/redhat-operators-6rt2d" Feb 25 13:45:37 crc kubenswrapper[4815]: I0225 13:45:37.571227 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z6fgs" podStartSLOduration=3.33671021 podStartE2EDuration="8.571205902s" podCreationTimestamp="2026-02-25 13:45:29 +0000 UTC" firstStartedPulling="2026-02-25 13:45:31.415330411 +0000 UTC m=+1489.216428475" lastFinishedPulling="2026-02-25 13:45:36.649826103 +0000 UTC m=+1494.450924167" observedRunningTime="2026-02-25 13:45:37.562212964 +0000 UTC m=+1495.363311018" watchObservedRunningTime="2026-02-25 13:45:37.571205902 +0000 UTC m=+1495.372303946" Feb 25 13:45:37 crc kubenswrapper[4815]: I0225 13:45:37.648629 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a84a587-7d0c-4201-b618-2e2caa068fbb-catalog-content\") pod \"redhat-operators-6rt2d\" (UID: \"2a84a587-7d0c-4201-b618-2e2caa068fbb\") " pod="openshift-marketplace/redhat-operators-6rt2d" Feb 25 13:45:37 crc kubenswrapper[4815]: I0225 13:45:37.649231 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a84a587-7d0c-4201-b618-2e2caa068fbb-catalog-content\") pod \"redhat-operators-6rt2d\" (UID: \"2a84a587-7d0c-4201-b618-2e2caa068fbb\") " pod="openshift-marketplace/redhat-operators-6rt2d" Feb 25 13:45:37 crc kubenswrapper[4815]: I0225 13:45:37.649799 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a84a587-7d0c-4201-b618-2e2caa068fbb-utilities\") pod \"redhat-operators-6rt2d\" (UID: \"2a84a587-7d0c-4201-b618-2e2caa068fbb\") " pod="openshift-marketplace/redhat-operators-6rt2d" Feb 25 13:45:37 crc kubenswrapper[4815]: I0225 13:45:37.650241 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a84a587-7d0c-4201-b618-2e2caa068fbb-utilities\") pod \"redhat-operators-6rt2d\" (UID: \"2a84a587-7d0c-4201-b618-2e2caa068fbb\") " pod="openshift-marketplace/redhat-operators-6rt2d" Feb 25 13:45:37 crc kubenswrapper[4815]: I0225 13:45:37.650324 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l54g8\" (UniqueName: \"kubernetes.io/projected/2a84a587-7d0c-4201-b618-2e2caa068fbb-kube-api-access-l54g8\") pod \"redhat-operators-6rt2d\" (UID: \"2a84a587-7d0c-4201-b618-2e2caa068fbb\") " pod="openshift-marketplace/redhat-operators-6rt2d" Feb 25 13:45:37 crc kubenswrapper[4815]: I0225 13:45:37.677589 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l54g8\" (UniqueName: \"kubernetes.io/projected/2a84a587-7d0c-4201-b618-2e2caa068fbb-kube-api-access-l54g8\") pod \"redhat-operators-6rt2d\" (UID: \"2a84a587-7d0c-4201-b618-2e2caa068fbb\") " pod="openshift-marketplace/redhat-operators-6rt2d" Feb 25 13:45:37 crc kubenswrapper[4815]: I0225 13:45:37.683199 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6rt2d" Feb 25 13:45:38 crc kubenswrapper[4815]: I0225 13:45:38.146340 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6rt2d"] Feb 25 13:45:38 crc kubenswrapper[4815]: W0225 13:45:38.155499 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a84a587_7d0c_4201_b618_2e2caa068fbb.slice/crio-57c6c796676db73e9e3709259314b91fa3f2547c54bc4edd3d6d13407047a426 WatchSource:0}: Error finding container 57c6c796676db73e9e3709259314b91fa3f2547c54bc4edd3d6d13407047a426: Status 404 returned error can't find the container with id 57c6c796676db73e9e3709259314b91fa3f2547c54bc4edd3d6d13407047a426 Feb 25 13:45:38 crc kubenswrapper[4815]: I0225 13:45:38.552327 4815 generic.go:334] "Generic (PLEG): container finished" podID="2a84a587-7d0c-4201-b618-2e2caa068fbb" containerID="f20a339cbcbbf0421d81b2f15a003fdb65bff7fd9dd65322979c607344cfc1b6" exitCode=0 Feb 25 13:45:38 crc kubenswrapper[4815]: I0225 13:45:38.552421 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rt2d" event={"ID":"2a84a587-7d0c-4201-b618-2e2caa068fbb","Type":"ContainerDied","Data":"f20a339cbcbbf0421d81b2f15a003fdb65bff7fd9dd65322979c607344cfc1b6"} Feb 25 13:45:38 crc kubenswrapper[4815]: I0225 13:45:38.552719 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rt2d" event={"ID":"2a84a587-7d0c-4201-b618-2e2caa068fbb","Type":"ContainerStarted","Data":"57c6c796676db73e9e3709259314b91fa3f2547c54bc4edd3d6d13407047a426"} Feb 25 13:45:39 crc kubenswrapper[4815]: I0225 13:45:39.895606 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-z6fgs" Feb 25 13:45:39 crc kubenswrapper[4815]: I0225 13:45:39.895915 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z6fgs" Feb 25 13:45:39 crc kubenswrapper[4815]: I0225 13:45:39.960301 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z6fgs" Feb 25 13:45:40 crc kubenswrapper[4815]: I0225 13:45:40.586701 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rt2d" event={"ID":"2a84a587-7d0c-4201-b618-2e2caa068fbb","Type":"ContainerStarted","Data":"2c3ff17a38beb9ec052510a27fcfbc89331d6ebd5fa1fdabd0e83e29ccd065b8"} Feb 25 13:45:42 crc kubenswrapper[4815]: I0225 13:45:42.609774 4815 generic.go:334] "Generic (PLEG): container finished" podID="2a84a587-7d0c-4201-b618-2e2caa068fbb" containerID="2c3ff17a38beb9ec052510a27fcfbc89331d6ebd5fa1fdabd0e83e29ccd065b8" exitCode=0 Feb 25 13:45:42 crc kubenswrapper[4815]: I0225 13:45:42.609860 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rt2d" event={"ID":"2a84a587-7d0c-4201-b618-2e2caa068fbb","Type":"ContainerDied","Data":"2c3ff17a38beb9ec052510a27fcfbc89331d6ebd5fa1fdabd0e83e29ccd065b8"} Feb 25 13:45:43 crc kubenswrapper[4815]: I0225 13:45:43.622458 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rt2d" event={"ID":"2a84a587-7d0c-4201-b618-2e2caa068fbb","Type":"ContainerStarted","Data":"6f3c5de272da935ecf97d48bf4a694fb80bf2dd98b4d7ac2f4e42069a678104f"} Feb 25 13:45:43 crc kubenswrapper[4815]: I0225 13:45:43.644586 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6rt2d" podStartSLOduration=2.182110487 podStartE2EDuration="6.644566546s" podCreationTimestamp="2026-02-25 13:45:37 +0000 UTC" firstStartedPulling="2026-02-25 13:45:38.554569409 +0000 UTC m=+1496.355667463" lastFinishedPulling="2026-02-25 13:45:43.017025468 +0000 UTC m=+1500.818123522" observedRunningTime="2026-02-25 13:45:43.638226711 +0000 UTC m=+1501.439324765" watchObservedRunningTime="2026-02-25 13:45:43.644566546 +0000 UTC m=+1501.445664600" Feb 25 13:45:45 crc kubenswrapper[4815]: I0225 13:45:45.827126 4815 scope.go:117] "RemoveContainer" containerID="c193180270c2be6a48f10c187ca9589ce0589e365ef03696be93dcd22dd25bb4" Feb 25 13:45:47 crc kubenswrapper[4815]: I0225 13:45:47.684336 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6rt2d" Feb 25 13:45:47 crc kubenswrapper[4815]: I0225 13:45:47.685723 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6rt2d" Feb 25 13:45:48 crc kubenswrapper[4815]: I0225 13:45:48.729021 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6rt2d" podUID="2a84a587-7d0c-4201-b618-2e2caa068fbb" containerName="registry-server" probeResult="failure" output=< Feb 25 13:45:48 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Feb 25 13:45:48 crc kubenswrapper[4815]: > Feb 25 13:45:49 crc kubenswrapper[4815]: I0225 13:45:49.956062 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z6fgs" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.046456 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z6fgs"] Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.097665 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mvxkh"] Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.098073 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mvxkh" podUID="7fae8512-7a85-4d34-87f3-d4d2e6832d93" containerName="registry-server" containerID="cri-o://addc89334739906dfdeed0b3362af2d90ea3537c14f3d8cf7cd4a20c2d24ea52" gracePeriod=2 Feb 25 13:45:50 crc kubenswrapper[4815]: E0225 13:45:50.376727 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fae8512_7a85_4d34_87f3_d4d2e6832d93.slice/crio-addc89334739906dfdeed0b3362af2d90ea3537c14f3d8cf7cd4a20c2d24ea52.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fae8512_7a85_4d34_87f3_d4d2e6832d93.slice/crio-conmon-addc89334739906dfdeed0b3362af2d90ea3537c14f3d8cf7cd4a20c2d24ea52.scope\": RecentStats: unable to find data in memory cache]" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.636249 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvxkh" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.679120 4815 generic.go:334] "Generic (PLEG): container finished" podID="7fae8512-7a85-4d34-87f3-d4d2e6832d93" containerID="addc89334739906dfdeed0b3362af2d90ea3537c14f3d8cf7cd4a20c2d24ea52" exitCode=0 Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.680029 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvxkh" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.680533 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvxkh" event={"ID":"7fae8512-7a85-4d34-87f3-d4d2e6832d93","Type":"ContainerDied","Data":"addc89334739906dfdeed0b3362af2d90ea3537c14f3d8cf7cd4a20c2d24ea52"} Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.680561 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvxkh" event={"ID":"7fae8512-7a85-4d34-87f3-d4d2e6832d93","Type":"ContainerDied","Data":"eb15f7573a805fe3c98ed29843d8799a0c5ffbdec3536975882964e7b51ab0b1"} Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.680578 4815 scope.go:117] "RemoveContainer" containerID="addc89334739906dfdeed0b3362af2d90ea3537c14f3d8cf7cd4a20c2d24ea52" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.712799 4815 scope.go:117] "RemoveContainer" containerID="2895998dbd5e494f739e100bda13cd2b777b28eec6f42ee38dc5f44c4baadf38" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.767819 4815 scope.go:117] "RemoveContainer" containerID="06afe600018f51c39a204ead37fdd6a4e0440a72377785d57ecaf1458e955962" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.811189 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fae8512-7a85-4d34-87f3-d4d2e6832d93-catalog-content\") pod \"7fae8512-7a85-4d34-87f3-d4d2e6832d93\" (UID: \"7fae8512-7a85-4d34-87f3-d4d2e6832d93\") " Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.811251 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fae8512-7a85-4d34-87f3-d4d2e6832d93-utilities\") pod \"7fae8512-7a85-4d34-87f3-d4d2e6832d93\" (UID: \"7fae8512-7a85-4d34-87f3-d4d2e6832d93\") " Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.811385 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldjqw\" (UniqueName: \"kubernetes.io/projected/7fae8512-7a85-4d34-87f3-d4d2e6832d93-kube-api-access-ldjqw\") pod \"7fae8512-7a85-4d34-87f3-d4d2e6832d93\" (UID: \"7fae8512-7a85-4d34-87f3-d4d2e6832d93\") " Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.813337 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fae8512-7a85-4d34-87f3-d4d2e6832d93-utilities" (OuterVolumeSpecName: "utilities") pod "7fae8512-7a85-4d34-87f3-d4d2e6832d93" (UID: "7fae8512-7a85-4d34-87f3-d4d2e6832d93"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.822152 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fae8512-7a85-4d34-87f3-d4d2e6832d93-kube-api-access-ldjqw" (OuterVolumeSpecName: "kube-api-access-ldjqw") pod "7fae8512-7a85-4d34-87f3-d4d2e6832d93" (UID: "7fae8512-7a85-4d34-87f3-d4d2e6832d93"). InnerVolumeSpecName "kube-api-access-ldjqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.836858 4815 scope.go:117] "RemoveContainer" containerID="addc89334739906dfdeed0b3362af2d90ea3537c14f3d8cf7cd4a20c2d24ea52" Feb 25 13:45:50 crc kubenswrapper[4815]: E0225 13:45:50.837413 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"addc89334739906dfdeed0b3362af2d90ea3537c14f3d8cf7cd4a20c2d24ea52\": container with ID starting with addc89334739906dfdeed0b3362af2d90ea3537c14f3d8cf7cd4a20c2d24ea52 not found: ID does not exist" containerID="addc89334739906dfdeed0b3362af2d90ea3537c14f3d8cf7cd4a20c2d24ea52" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.837463 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"addc89334739906dfdeed0b3362af2d90ea3537c14f3d8cf7cd4a20c2d24ea52"} err="failed to get container status \"addc89334739906dfdeed0b3362af2d90ea3537c14f3d8cf7cd4a20c2d24ea52\": rpc error: code = NotFound desc = could not find container \"addc89334739906dfdeed0b3362af2d90ea3537c14f3d8cf7cd4a20c2d24ea52\": container with ID starting with addc89334739906dfdeed0b3362af2d90ea3537c14f3d8cf7cd4a20c2d24ea52 not found: ID does not exist" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.837491 4815 scope.go:117] "RemoveContainer" containerID="2895998dbd5e494f739e100bda13cd2b777b28eec6f42ee38dc5f44c4baadf38" Feb 25 13:45:50 crc kubenswrapper[4815]: E0225 13:45:50.837801 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2895998dbd5e494f739e100bda13cd2b777b28eec6f42ee38dc5f44c4baadf38\": container with ID starting with 2895998dbd5e494f739e100bda13cd2b777b28eec6f42ee38dc5f44c4baadf38 not found: ID does not exist" containerID="2895998dbd5e494f739e100bda13cd2b777b28eec6f42ee38dc5f44c4baadf38" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.837825 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2895998dbd5e494f739e100bda13cd2b777b28eec6f42ee38dc5f44c4baadf38"} err="failed to get container status \"2895998dbd5e494f739e100bda13cd2b777b28eec6f42ee38dc5f44c4baadf38\": rpc error: code = NotFound desc = could not find container \"2895998dbd5e494f739e100bda13cd2b777b28eec6f42ee38dc5f44c4baadf38\": container with ID starting with 2895998dbd5e494f739e100bda13cd2b777b28eec6f42ee38dc5f44c4baadf38 not found: ID does not exist" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.837842 4815 scope.go:117] "RemoveContainer" containerID="06afe600018f51c39a204ead37fdd6a4e0440a72377785d57ecaf1458e955962" Feb 25 13:45:50 crc kubenswrapper[4815]: E0225 13:45:50.838372 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06afe600018f51c39a204ead37fdd6a4e0440a72377785d57ecaf1458e955962\": container with ID starting with 06afe600018f51c39a204ead37fdd6a4e0440a72377785d57ecaf1458e955962 not found: ID does not exist" containerID="06afe600018f51c39a204ead37fdd6a4e0440a72377785d57ecaf1458e955962" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.838396 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06afe600018f51c39a204ead37fdd6a4e0440a72377785d57ecaf1458e955962"} err="failed to get container status \"06afe600018f51c39a204ead37fdd6a4e0440a72377785d57ecaf1458e955962\": rpc error: code = NotFound desc = could not find container \"06afe600018f51c39a204ead37fdd6a4e0440a72377785d57ecaf1458e955962\": container with ID starting with 06afe600018f51c39a204ead37fdd6a4e0440a72377785d57ecaf1458e955962 not found: ID does not exist" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.878137 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fae8512-7a85-4d34-87f3-d4d2e6832d93-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fae8512-7a85-4d34-87f3-d4d2e6832d93" (UID: "7fae8512-7a85-4d34-87f3-d4d2e6832d93"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.914224 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldjqw\" (UniqueName: \"kubernetes.io/projected/7fae8512-7a85-4d34-87f3-d4d2e6832d93-kube-api-access-ldjqw\") on node \"crc\" DevicePath \"\"" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.914257 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fae8512-7a85-4d34-87f3-d4d2e6832d93-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:45:50 crc kubenswrapper[4815]: I0225 13:45:50.914266 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fae8512-7a85-4d34-87f3-d4d2e6832d93-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:45:51 crc kubenswrapper[4815]: I0225 13:45:51.004646 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mvxkh"] Feb 25 13:45:51 crc kubenswrapper[4815]: I0225 13:45:51.013187 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mvxkh"] Feb 25 13:45:51 crc kubenswrapper[4815]: I0225 13:45:51.342137 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:45:51 crc kubenswrapper[4815]: I0225 13:45:51.342475 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:45:52 crc kubenswrapper[4815]: I0225 13:45:52.949679 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fae8512-7a85-4d34-87f3-d4d2e6832d93" path="/var/lib/kubelet/pods/7fae8512-7a85-4d34-87f3-d4d2e6832d93/volumes" Feb 25 13:45:57 crc kubenswrapper[4815]: I0225 13:45:57.761532 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6rt2d" Feb 25 13:45:57 crc kubenswrapper[4815]: I0225 13:45:57.825319 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6rt2d" Feb 25 13:45:59 crc kubenswrapper[4815]: I0225 13:45:59.015133 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6rt2d"] Feb 25 13:45:59 crc kubenswrapper[4815]: I0225 13:45:59.759881 4815 generic.go:334] "Generic (PLEG): container finished" podID="3a6985a9-1674-41ff-b441-23a83eff2148" containerID="0436aaf0dbd5a6ae739f8097cf700105a672b8b981562794d8114146df3db7eb" exitCode=0 Feb 25 13:45:59 crc kubenswrapper[4815]: I0225 13:45:59.759966 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" event={"ID":"3a6985a9-1674-41ff-b441-23a83eff2148","Type":"ContainerDied","Data":"0436aaf0dbd5a6ae739f8097cf700105a672b8b981562794d8114146df3db7eb"} Feb 25 13:45:59 crc kubenswrapper[4815]: I0225 13:45:59.760135 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6rt2d" podUID="2a84a587-7d0c-4201-b618-2e2caa068fbb" containerName="registry-server" containerID="cri-o://6f3c5de272da935ecf97d48bf4a694fb80bf2dd98b4d7ac2f4e42069a678104f" gracePeriod=2 Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.147015 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533786-j5c6h"] Feb 25 13:46:00 crc kubenswrapper[4815]: E0225 13:46:00.147915 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fae8512-7a85-4d34-87f3-d4d2e6832d93" containerName="extract-utilities" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.147933 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fae8512-7a85-4d34-87f3-d4d2e6832d93" containerName="extract-utilities" Feb 25 13:46:00 crc kubenswrapper[4815]: E0225 13:46:00.147952 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fae8512-7a85-4d34-87f3-d4d2e6832d93" containerName="extract-content" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.147960 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fae8512-7a85-4d34-87f3-d4d2e6832d93" containerName="extract-content" Feb 25 13:46:00 crc kubenswrapper[4815]: E0225 13:46:00.147980 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fae8512-7a85-4d34-87f3-d4d2e6832d93" containerName="registry-server" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.147987 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fae8512-7a85-4d34-87f3-d4d2e6832d93" containerName="registry-server" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.148209 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fae8512-7a85-4d34-87f3-d4d2e6832d93" containerName="registry-server" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.149137 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533786-j5c6h" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.151876 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.153435 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.156903 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.168759 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533786-j5c6h"] Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.252630 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6rt2d" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.298271 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bdrj\" (UniqueName: \"kubernetes.io/projected/657b0b94-e86a-487a-98e6-54f445ccab68-kube-api-access-7bdrj\") pod \"auto-csr-approver-29533786-j5c6h\" (UID: \"657b0b94-e86a-487a-98e6-54f445ccab68\") " pod="openshift-infra/auto-csr-approver-29533786-j5c6h" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.399880 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a84a587-7d0c-4201-b618-2e2caa068fbb-catalog-content\") pod \"2a84a587-7d0c-4201-b618-2e2caa068fbb\" (UID: \"2a84a587-7d0c-4201-b618-2e2caa068fbb\") " Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.399982 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l54g8\" (UniqueName: \"kubernetes.io/projected/2a84a587-7d0c-4201-b618-2e2caa068fbb-kube-api-access-l54g8\") pod \"2a84a587-7d0c-4201-b618-2e2caa068fbb\" (UID: \"2a84a587-7d0c-4201-b618-2e2caa068fbb\") " Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.400023 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a84a587-7d0c-4201-b618-2e2caa068fbb-utilities\") pod \"2a84a587-7d0c-4201-b618-2e2caa068fbb\" (UID: \"2a84a587-7d0c-4201-b618-2e2caa068fbb\") " Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.400774 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bdrj\" (UniqueName: \"kubernetes.io/projected/657b0b94-e86a-487a-98e6-54f445ccab68-kube-api-access-7bdrj\") pod \"auto-csr-approver-29533786-j5c6h\" (UID: \"657b0b94-e86a-487a-98e6-54f445ccab68\") " pod="openshift-infra/auto-csr-approver-29533786-j5c6h" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.401551 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a84a587-7d0c-4201-b618-2e2caa068fbb-utilities" (OuterVolumeSpecName: "utilities") pod "2a84a587-7d0c-4201-b618-2e2caa068fbb" (UID: "2a84a587-7d0c-4201-b618-2e2caa068fbb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.414912 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a84a587-7d0c-4201-b618-2e2caa068fbb-kube-api-access-l54g8" (OuterVolumeSpecName: "kube-api-access-l54g8") pod "2a84a587-7d0c-4201-b618-2e2caa068fbb" (UID: "2a84a587-7d0c-4201-b618-2e2caa068fbb"). InnerVolumeSpecName "kube-api-access-l54g8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.419192 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bdrj\" (UniqueName: \"kubernetes.io/projected/657b0b94-e86a-487a-98e6-54f445ccab68-kube-api-access-7bdrj\") pod \"auto-csr-approver-29533786-j5c6h\" (UID: \"657b0b94-e86a-487a-98e6-54f445ccab68\") " pod="openshift-infra/auto-csr-approver-29533786-j5c6h" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.517974 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l54g8\" (UniqueName: \"kubernetes.io/projected/2a84a587-7d0c-4201-b618-2e2caa068fbb-kube-api-access-l54g8\") on node \"crc\" DevicePath \"\"" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.518047 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a84a587-7d0c-4201-b618-2e2caa068fbb-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.548683 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533786-j5c6h" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.573353 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a84a587-7d0c-4201-b618-2e2caa068fbb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a84a587-7d0c-4201-b618-2e2caa068fbb" (UID: "2a84a587-7d0c-4201-b618-2e2caa068fbb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.621557 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a84a587-7d0c-4201-b618-2e2caa068fbb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.770098 4815 generic.go:334] "Generic (PLEG): container finished" podID="2a84a587-7d0c-4201-b618-2e2caa068fbb" containerID="6f3c5de272da935ecf97d48bf4a694fb80bf2dd98b4d7ac2f4e42069a678104f" exitCode=0 Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.770146 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rt2d" event={"ID":"2a84a587-7d0c-4201-b618-2e2caa068fbb","Type":"ContainerDied","Data":"6f3c5de272da935ecf97d48bf4a694fb80bf2dd98b4d7ac2f4e42069a678104f"} Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.770668 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rt2d" event={"ID":"2a84a587-7d0c-4201-b618-2e2caa068fbb","Type":"ContainerDied","Data":"57c6c796676db73e9e3709259314b91fa3f2547c54bc4edd3d6d13407047a426"} Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.770692 4815 scope.go:117] "RemoveContainer" containerID="6f3c5de272da935ecf97d48bf4a694fb80bf2dd98b4d7ac2f4e42069a678104f" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.770218 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6rt2d" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.799424 4815 scope.go:117] "RemoveContainer" containerID="2c3ff17a38beb9ec052510a27fcfbc89331d6ebd5fa1fdabd0e83e29ccd065b8" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.826976 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6rt2d"] Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.840826 4815 scope.go:117] "RemoveContainer" containerID="f20a339cbcbbf0421d81b2f15a003fdb65bff7fd9dd65322979c607344cfc1b6" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.842820 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6rt2d"] Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.864734 4815 scope.go:117] "RemoveContainer" containerID="6f3c5de272da935ecf97d48bf4a694fb80bf2dd98b4d7ac2f4e42069a678104f" Feb 25 13:46:00 crc kubenswrapper[4815]: E0225 13:46:00.865912 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f3c5de272da935ecf97d48bf4a694fb80bf2dd98b4d7ac2f4e42069a678104f\": container with ID starting with 6f3c5de272da935ecf97d48bf4a694fb80bf2dd98b4d7ac2f4e42069a678104f not found: ID does not exist" containerID="6f3c5de272da935ecf97d48bf4a694fb80bf2dd98b4d7ac2f4e42069a678104f" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.865942 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f3c5de272da935ecf97d48bf4a694fb80bf2dd98b4d7ac2f4e42069a678104f"} err="failed to get container status \"6f3c5de272da935ecf97d48bf4a694fb80bf2dd98b4d7ac2f4e42069a678104f\": rpc error: code = NotFound desc = could not find container \"6f3c5de272da935ecf97d48bf4a694fb80bf2dd98b4d7ac2f4e42069a678104f\": container with ID starting with 6f3c5de272da935ecf97d48bf4a694fb80bf2dd98b4d7ac2f4e42069a678104f not found: ID does not exist" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.865969 4815 scope.go:117] "RemoveContainer" containerID="2c3ff17a38beb9ec052510a27fcfbc89331d6ebd5fa1fdabd0e83e29ccd065b8" Feb 25 13:46:00 crc kubenswrapper[4815]: E0225 13:46:00.870952 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c3ff17a38beb9ec052510a27fcfbc89331d6ebd5fa1fdabd0e83e29ccd065b8\": container with ID starting with 2c3ff17a38beb9ec052510a27fcfbc89331d6ebd5fa1fdabd0e83e29ccd065b8 not found: ID does not exist" containerID="2c3ff17a38beb9ec052510a27fcfbc89331d6ebd5fa1fdabd0e83e29ccd065b8" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.871013 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c3ff17a38beb9ec052510a27fcfbc89331d6ebd5fa1fdabd0e83e29ccd065b8"} err="failed to get container status \"2c3ff17a38beb9ec052510a27fcfbc89331d6ebd5fa1fdabd0e83e29ccd065b8\": rpc error: code = NotFound desc = could not find container \"2c3ff17a38beb9ec052510a27fcfbc89331d6ebd5fa1fdabd0e83e29ccd065b8\": container with ID starting with 2c3ff17a38beb9ec052510a27fcfbc89331d6ebd5fa1fdabd0e83e29ccd065b8 not found: ID does not exist" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.871046 4815 scope.go:117] "RemoveContainer" containerID="f20a339cbcbbf0421d81b2f15a003fdb65bff7fd9dd65322979c607344cfc1b6" Feb 25 13:46:00 crc kubenswrapper[4815]: E0225 13:46:00.871481 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f20a339cbcbbf0421d81b2f15a003fdb65bff7fd9dd65322979c607344cfc1b6\": container with ID starting with f20a339cbcbbf0421d81b2f15a003fdb65bff7fd9dd65322979c607344cfc1b6 not found: ID does not exist" containerID="f20a339cbcbbf0421d81b2f15a003fdb65bff7fd9dd65322979c607344cfc1b6" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.871535 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f20a339cbcbbf0421d81b2f15a003fdb65bff7fd9dd65322979c607344cfc1b6"} err="failed to get container status \"f20a339cbcbbf0421d81b2f15a003fdb65bff7fd9dd65322979c607344cfc1b6\": rpc error: code = NotFound desc = could not find container \"f20a339cbcbbf0421d81b2f15a003fdb65bff7fd9dd65322979c607344cfc1b6\": container with ID starting with f20a339cbcbbf0421d81b2f15a003fdb65bff7fd9dd65322979c607344cfc1b6 not found: ID does not exist" Feb 25 13:46:00 crc kubenswrapper[4815]: I0225 13:46:00.946412 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a84a587-7d0c-4201-b618-2e2caa068fbb" path="/var/lib/kubelet/pods/2a84a587-7d0c-4201-b618-2e2caa068fbb/volumes" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:00.991352 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533786-j5c6h"] Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.240232 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.340619 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-ssh-key-openstack-edpm-ipam\") pod \"3a6985a9-1674-41ff-b441-23a83eff2148\" (UID: \"3a6985a9-1674-41ff-b441-23a83eff2148\") " Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.340754 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-bootstrap-combined-ca-bundle\") pod \"3a6985a9-1674-41ff-b441-23a83eff2148\" (UID: \"3a6985a9-1674-41ff-b441-23a83eff2148\") " Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.340809 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-inventory\") pod \"3a6985a9-1674-41ff-b441-23a83eff2148\" (UID: \"3a6985a9-1674-41ff-b441-23a83eff2148\") " Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.340866 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5fkw\" (UniqueName: \"kubernetes.io/projected/3a6985a9-1674-41ff-b441-23a83eff2148-kube-api-access-j5fkw\") pod \"3a6985a9-1674-41ff-b441-23a83eff2148\" (UID: \"3a6985a9-1674-41ff-b441-23a83eff2148\") " Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.347147 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3a6985a9-1674-41ff-b441-23a83eff2148" (UID: "3a6985a9-1674-41ff-b441-23a83eff2148"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.347938 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a6985a9-1674-41ff-b441-23a83eff2148-kube-api-access-j5fkw" (OuterVolumeSpecName: "kube-api-access-j5fkw") pod "3a6985a9-1674-41ff-b441-23a83eff2148" (UID: "3a6985a9-1674-41ff-b441-23a83eff2148"). InnerVolumeSpecName "kube-api-access-j5fkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.374388 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3a6985a9-1674-41ff-b441-23a83eff2148" (UID: "3a6985a9-1674-41ff-b441-23a83eff2148"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.375266 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-inventory" (OuterVolumeSpecName: "inventory") pod "3a6985a9-1674-41ff-b441-23a83eff2148" (UID: "3a6985a9-1674-41ff-b441-23a83eff2148"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.442685 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.442725 4815 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.442744 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a6985a9-1674-41ff-b441-23a83eff2148-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.442758 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5fkw\" (UniqueName: \"kubernetes.io/projected/3a6985a9-1674-41ff-b441-23a83eff2148-kube-api-access-j5fkw\") on node \"crc\" DevicePath \"\"" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.788764 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533786-j5c6h" event={"ID":"657b0b94-e86a-487a-98e6-54f445ccab68","Type":"ContainerStarted","Data":"33e0ffb40151fa2449a4ee1015e2114a3900c38cef06b67e1aebb860035ae12a"} Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.791467 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" event={"ID":"3a6985a9-1674-41ff-b441-23a83eff2148","Type":"ContainerDied","Data":"77a68cee2e5c93077335c90691088506eecd6d98cc45531d78e6d3c272e5f36a"} Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.791536 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77a68cee2e5c93077335c90691088506eecd6d98cc45531d78e6d3c272e5f36a" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.791549 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.875580 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v"] Feb 25 13:46:01 crc kubenswrapper[4815]: E0225 13:46:01.876182 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a84a587-7d0c-4201-b618-2e2caa068fbb" containerName="extract-content" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.876201 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a84a587-7d0c-4201-b618-2e2caa068fbb" containerName="extract-content" Feb 25 13:46:01 crc kubenswrapper[4815]: E0225 13:46:01.876224 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a84a587-7d0c-4201-b618-2e2caa068fbb" containerName="extract-utilities" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.876235 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a84a587-7d0c-4201-b618-2e2caa068fbb" containerName="extract-utilities" Feb 25 13:46:01 crc kubenswrapper[4815]: E0225 13:46:01.876259 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a84a587-7d0c-4201-b618-2e2caa068fbb" containerName="registry-server" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.876268 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a84a587-7d0c-4201-b618-2e2caa068fbb" containerName="registry-server" Feb 25 13:46:01 crc kubenswrapper[4815]: E0225 13:46:01.876283 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a6985a9-1674-41ff-b441-23a83eff2148" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.876295 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a6985a9-1674-41ff-b441-23a83eff2148" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.876545 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a6985a9-1674-41ff-b441-23a83eff2148" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.876576 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a84a587-7d0c-4201-b618-2e2caa068fbb" containerName="registry-server" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.877439 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.880296 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mmbj" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.880620 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.882974 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.883272 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.895340 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v"] Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.954789 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v\" (UID: \"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.955175 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v\" (UID: \"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" Feb 25 13:46:01 crc kubenswrapper[4815]: I0225 13:46:01.955247 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r4jw\" (UniqueName: \"kubernetes.io/projected/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-kube-api-access-5r4jw\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v\" (UID: \"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" Feb 25 13:46:02 crc kubenswrapper[4815]: I0225 13:46:02.057013 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v\" (UID: \"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" Feb 25 13:46:02 crc kubenswrapper[4815]: I0225 13:46:02.057489 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r4jw\" (UniqueName: \"kubernetes.io/projected/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-kube-api-access-5r4jw\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v\" (UID: \"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" Feb 25 13:46:02 crc kubenswrapper[4815]: I0225 13:46:02.057700 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v\" (UID: \"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" Feb 25 13:46:02 crc kubenswrapper[4815]: I0225 13:46:02.064740 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v\" (UID: \"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" Feb 25 13:46:02 crc kubenswrapper[4815]: I0225 13:46:02.064767 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v\" (UID: \"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" Feb 25 13:46:02 crc kubenswrapper[4815]: I0225 13:46:02.075988 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r4jw\" (UniqueName: \"kubernetes.io/projected/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-kube-api-access-5r4jw\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v\" (UID: \"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" Feb 25 13:46:02 crc kubenswrapper[4815]: I0225 13:46:02.234786 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" Feb 25 13:46:02 crc kubenswrapper[4815]: I0225 13:46:02.753132 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v"] Feb 25 13:46:02 crc kubenswrapper[4815]: W0225 13:46:02.760242 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d6402d2_d86e_4529_a4bb_b6b26f9ed63c.slice/crio-c05dd6dc43365887ea0824b060077adc4109b377a4b77904e619dec4e6ea01de WatchSource:0}: Error finding container c05dd6dc43365887ea0824b060077adc4109b377a4b77904e619dec4e6ea01de: Status 404 returned error can't find the container with id c05dd6dc43365887ea0824b060077adc4109b377a4b77904e619dec4e6ea01de Feb 25 13:46:02 crc kubenswrapper[4815]: I0225 13:46:02.805349 4815 generic.go:334] "Generic (PLEG): container finished" podID="657b0b94-e86a-487a-98e6-54f445ccab68" containerID="c3ec72f599c5ebb4e95c902f08a9d31ba8859ee74bfe21b74301d6bd884f3850" exitCode=0 Feb 25 13:46:02 crc kubenswrapper[4815]: I0225 13:46:02.805434 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533786-j5c6h" event={"ID":"657b0b94-e86a-487a-98e6-54f445ccab68","Type":"ContainerDied","Data":"c3ec72f599c5ebb4e95c902f08a9d31ba8859ee74bfe21b74301d6bd884f3850"} Feb 25 13:46:02 crc kubenswrapper[4815]: I0225 13:46:02.807433 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" event={"ID":"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c","Type":"ContainerStarted","Data":"c05dd6dc43365887ea0824b060077adc4109b377a4b77904e619dec4e6ea01de"} Feb 25 13:46:03 crc kubenswrapper[4815]: I0225 13:46:03.817262 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" event={"ID":"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c","Type":"ContainerStarted","Data":"69873e4101a7a7278d6103f8c8fc828b6863b9922a48d96293639d50396b59a3"} Feb 25 13:46:03 crc kubenswrapper[4815]: I0225 13:46:03.839976 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" podStartSLOduration=2.436243361 podStartE2EDuration="2.839945692s" podCreationTimestamp="2026-02-25 13:46:01 +0000 UTC" firstStartedPulling="2026-02-25 13:46:02.76627667 +0000 UTC m=+1520.567374724" lastFinishedPulling="2026-02-25 13:46:03.169979001 +0000 UTC m=+1520.971077055" observedRunningTime="2026-02-25 13:46:03.837809755 +0000 UTC m=+1521.638907809" watchObservedRunningTime="2026-02-25 13:46:03.839945692 +0000 UTC m=+1521.641043756" Feb 25 13:46:04 crc kubenswrapper[4815]: I0225 13:46:04.231416 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533786-j5c6h" Feb 25 13:46:04 crc kubenswrapper[4815]: I0225 13:46:04.301693 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bdrj\" (UniqueName: \"kubernetes.io/projected/657b0b94-e86a-487a-98e6-54f445ccab68-kube-api-access-7bdrj\") pod \"657b0b94-e86a-487a-98e6-54f445ccab68\" (UID: \"657b0b94-e86a-487a-98e6-54f445ccab68\") " Feb 25 13:46:04 crc kubenswrapper[4815]: I0225 13:46:04.306418 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/657b0b94-e86a-487a-98e6-54f445ccab68-kube-api-access-7bdrj" (OuterVolumeSpecName: "kube-api-access-7bdrj") pod "657b0b94-e86a-487a-98e6-54f445ccab68" (UID: "657b0b94-e86a-487a-98e6-54f445ccab68"). InnerVolumeSpecName "kube-api-access-7bdrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:46:04 crc kubenswrapper[4815]: I0225 13:46:04.403748 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bdrj\" (UniqueName: \"kubernetes.io/projected/657b0b94-e86a-487a-98e6-54f445ccab68-kube-api-access-7bdrj\") on node \"crc\" DevicePath \"\"" Feb 25 13:46:04 crc kubenswrapper[4815]: I0225 13:46:04.828787 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533786-j5c6h" Feb 25 13:46:04 crc kubenswrapper[4815]: I0225 13:46:04.834642 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533786-j5c6h" event={"ID":"657b0b94-e86a-487a-98e6-54f445ccab68","Type":"ContainerDied","Data":"33e0ffb40151fa2449a4ee1015e2114a3900c38cef06b67e1aebb860035ae12a"} Feb 25 13:46:04 crc kubenswrapper[4815]: I0225 13:46:04.834678 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33e0ffb40151fa2449a4ee1015e2114a3900c38cef06b67e1aebb860035ae12a" Feb 25 13:46:05 crc kubenswrapper[4815]: I0225 13:46:05.307544 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533780-vbrwb"] Feb 25 13:46:05 crc kubenswrapper[4815]: I0225 13:46:05.318556 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533780-vbrwb"] Feb 25 13:46:06 crc kubenswrapper[4815]: I0225 13:46:06.956394 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a63b31e2-6a8c-4664-a47c-30f2a731c8a4" path="/var/lib/kubelet/pods/a63b31e2-6a8c-4664-a47c-30f2a731c8a4/volumes" Feb 25 13:46:12 crc kubenswrapper[4815]: I0225 13:46:12.793455 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dsjpl"] Feb 25 13:46:12 crc kubenswrapper[4815]: E0225 13:46:12.794471 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="657b0b94-e86a-487a-98e6-54f445ccab68" containerName="oc" Feb 25 13:46:12 crc kubenswrapper[4815]: I0225 13:46:12.794487 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="657b0b94-e86a-487a-98e6-54f445ccab68" containerName="oc" Feb 25 13:46:12 crc kubenswrapper[4815]: I0225 13:46:12.794697 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="657b0b94-e86a-487a-98e6-54f445ccab68" containerName="oc" Feb 25 13:46:12 crc kubenswrapper[4815]: I0225 13:46:12.795959 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dsjpl" Feb 25 13:46:12 crc kubenswrapper[4815]: I0225 13:46:12.811250 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dsjpl"] Feb 25 13:46:12 crc kubenswrapper[4815]: I0225 13:46:12.904173 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn8zs\" (UniqueName: \"kubernetes.io/projected/0781a2e4-ca50-4a4a-b35d-d4e448e99f9c-kube-api-access-kn8zs\") pod \"certified-operators-dsjpl\" (UID: \"0781a2e4-ca50-4a4a-b35d-d4e448e99f9c\") " pod="openshift-marketplace/certified-operators-dsjpl" Feb 25 13:46:12 crc kubenswrapper[4815]: I0225 13:46:12.904911 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0781a2e4-ca50-4a4a-b35d-d4e448e99f9c-utilities\") pod \"certified-operators-dsjpl\" (UID: \"0781a2e4-ca50-4a4a-b35d-d4e448e99f9c\") " pod="openshift-marketplace/certified-operators-dsjpl" Feb 25 13:46:12 crc kubenswrapper[4815]: I0225 13:46:12.905042 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0781a2e4-ca50-4a4a-b35d-d4e448e99f9c-catalog-content\") pod \"certified-operators-dsjpl\" (UID: \"0781a2e4-ca50-4a4a-b35d-d4e448e99f9c\") " pod="openshift-marketplace/certified-operators-dsjpl" Feb 25 13:46:13 crc kubenswrapper[4815]: I0225 13:46:13.007522 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn8zs\" (UniqueName: \"kubernetes.io/projected/0781a2e4-ca50-4a4a-b35d-d4e448e99f9c-kube-api-access-kn8zs\") pod \"certified-operators-dsjpl\" (UID: \"0781a2e4-ca50-4a4a-b35d-d4e448e99f9c\") " pod="openshift-marketplace/certified-operators-dsjpl" Feb 25 13:46:13 crc kubenswrapper[4815]: I0225 13:46:13.007650 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0781a2e4-ca50-4a4a-b35d-d4e448e99f9c-utilities\") pod \"certified-operators-dsjpl\" (UID: \"0781a2e4-ca50-4a4a-b35d-d4e448e99f9c\") " pod="openshift-marketplace/certified-operators-dsjpl" Feb 25 13:46:13 crc kubenswrapper[4815]: I0225 13:46:13.007687 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0781a2e4-ca50-4a4a-b35d-d4e448e99f9c-catalog-content\") pod \"certified-operators-dsjpl\" (UID: \"0781a2e4-ca50-4a4a-b35d-d4e448e99f9c\") " pod="openshift-marketplace/certified-operators-dsjpl" Feb 25 13:46:13 crc kubenswrapper[4815]: I0225 13:46:13.008148 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0781a2e4-ca50-4a4a-b35d-d4e448e99f9c-utilities\") pod \"certified-operators-dsjpl\" (UID: \"0781a2e4-ca50-4a4a-b35d-d4e448e99f9c\") " pod="openshift-marketplace/certified-operators-dsjpl" Feb 25 13:46:13 crc kubenswrapper[4815]: I0225 13:46:13.008263 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0781a2e4-ca50-4a4a-b35d-d4e448e99f9c-catalog-content\") pod \"certified-operators-dsjpl\" (UID: \"0781a2e4-ca50-4a4a-b35d-d4e448e99f9c\") " pod="openshift-marketplace/certified-operators-dsjpl" Feb 25 13:46:13 crc kubenswrapper[4815]: I0225 13:46:13.037033 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn8zs\" (UniqueName: \"kubernetes.io/projected/0781a2e4-ca50-4a4a-b35d-d4e448e99f9c-kube-api-access-kn8zs\") pod \"certified-operators-dsjpl\" (UID: \"0781a2e4-ca50-4a4a-b35d-d4e448e99f9c\") " pod="openshift-marketplace/certified-operators-dsjpl" Feb 25 13:46:13 crc kubenswrapper[4815]: I0225 13:46:13.121858 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dsjpl" Feb 25 13:46:13 crc kubenswrapper[4815]: I0225 13:46:13.624094 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dsjpl"] Feb 25 13:46:13 crc kubenswrapper[4815]: I0225 13:46:13.917169 4815 generic.go:334] "Generic (PLEG): container finished" podID="0781a2e4-ca50-4a4a-b35d-d4e448e99f9c" containerID="37bde9ea36c4b20af3996e2e15dd7ac8034b7b04b4326bf912707b52e136d94a" exitCode=0 Feb 25 13:46:13 crc kubenswrapper[4815]: I0225 13:46:13.917253 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dsjpl" event={"ID":"0781a2e4-ca50-4a4a-b35d-d4e448e99f9c","Type":"ContainerDied","Data":"37bde9ea36c4b20af3996e2e15dd7ac8034b7b04b4326bf912707b52e136d94a"} Feb 25 13:46:13 crc kubenswrapper[4815]: I0225 13:46:13.917300 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dsjpl" event={"ID":"0781a2e4-ca50-4a4a-b35d-d4e448e99f9c","Type":"ContainerStarted","Data":"203466215a4faa00c368d55048e7ea57f174a99b59da25ad3fae0bd899b9547f"} Feb 25 13:46:18 crc kubenswrapper[4815]: I0225 13:46:18.972501 4815 generic.go:334] "Generic (PLEG): container finished" podID="0781a2e4-ca50-4a4a-b35d-d4e448e99f9c" containerID="caad12f786ddb6df8bbf9d5d052c57b8c39321c18375a4bede412b21cfd83891" exitCode=0 Feb 25 13:46:18 crc kubenswrapper[4815]: I0225 13:46:18.972590 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dsjpl" event={"ID":"0781a2e4-ca50-4a4a-b35d-d4e448e99f9c","Type":"ContainerDied","Data":"caad12f786ddb6df8bbf9d5d052c57b8c39321c18375a4bede412b21cfd83891"} Feb 25 13:46:19 crc kubenswrapper[4815]: I0225 13:46:19.984639 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dsjpl" event={"ID":"0781a2e4-ca50-4a4a-b35d-d4e448e99f9c","Type":"ContainerStarted","Data":"f2b0448a814adb00d3c2e6f1a9a660bf77dd5c3253f4b735ecdefad89bfb73a8"} Feb 25 13:46:20 crc kubenswrapper[4815]: I0225 13:46:20.013285 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dsjpl" podStartSLOduration=2.555623912 podStartE2EDuration="8.013268235s" podCreationTimestamp="2026-02-25 13:46:12 +0000 UTC" firstStartedPulling="2026-02-25 13:46:13.919136317 +0000 UTC m=+1531.720234371" lastFinishedPulling="2026-02-25 13:46:19.37678063 +0000 UTC m=+1537.177878694" observedRunningTime="2026-02-25 13:46:20.003399399 +0000 UTC m=+1537.804497463" watchObservedRunningTime="2026-02-25 13:46:20.013268235 +0000 UTC m=+1537.814366289" Feb 25 13:46:21 crc kubenswrapper[4815]: I0225 13:46:21.342560 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:46:21 crc kubenswrapper[4815]: I0225 13:46:21.342911 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:46:21 crc kubenswrapper[4815]: I0225 13:46:21.343008 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:46:21 crc kubenswrapper[4815]: I0225 13:46:21.344281 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b"} pod="openshift-machine-config-operator/machine-config-daemon-frc75" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 13:46:21 crc kubenswrapper[4815]: I0225 13:46:21.344375 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" containerID="cri-o://faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" gracePeriod=600 Feb 25 13:46:21 crc kubenswrapper[4815]: E0225 13:46:21.463668 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:46:22 crc kubenswrapper[4815]: I0225 13:46:22.010390 4815 generic.go:334] "Generic (PLEG): container finished" podID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" exitCode=0 Feb 25 13:46:22 crc kubenswrapper[4815]: I0225 13:46:22.010452 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerDied","Data":"faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b"} Feb 25 13:46:22 crc kubenswrapper[4815]: I0225 13:46:22.010518 4815 scope.go:117] "RemoveContainer" containerID="f21f73b0f5d16d6a34d568e8c21d1a82e30512997d6c1c659ca1e100017fbcdd" Feb 25 13:46:22 crc kubenswrapper[4815]: I0225 13:46:22.012112 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:46:22 crc kubenswrapper[4815]: E0225 13:46:22.012720 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:46:23 crc kubenswrapper[4815]: I0225 13:46:23.122597 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dsjpl" Feb 25 13:46:23 crc kubenswrapper[4815]: I0225 13:46:23.122694 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dsjpl" Feb 25 13:46:23 crc kubenswrapper[4815]: I0225 13:46:23.201481 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dsjpl" Feb 25 13:46:24 crc kubenswrapper[4815]: I0225 13:46:24.109300 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dsjpl" Feb 25 13:46:24 crc kubenswrapper[4815]: I0225 13:46:24.201078 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dsjpl"] Feb 25 13:46:24 crc kubenswrapper[4815]: I0225 13:46:24.263354 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b4jcd"] Feb 25 13:46:24 crc kubenswrapper[4815]: I0225 13:46:24.263722 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b4jcd" podUID="f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc" containerName="registry-server" containerID="cri-o://f94fd26d90a1ba7ce5481ebdce566b03cc4fa187919a2817bdad927c880a5288" gracePeriod=2 Feb 25 13:46:24 crc kubenswrapper[4815]: I0225 13:46:24.747134 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b4jcd" Feb 25 13:46:24 crc kubenswrapper[4815]: I0225 13:46:24.855750 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdxsz\" (UniqueName: \"kubernetes.io/projected/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-kube-api-access-sdxsz\") pod \"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc\" (UID: \"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc\") " Feb 25 13:46:24 crc kubenswrapper[4815]: I0225 13:46:24.855844 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-utilities\") pod \"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc\" (UID: \"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc\") " Feb 25 13:46:24 crc kubenswrapper[4815]: I0225 13:46:24.855964 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-catalog-content\") pod \"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc\" (UID: \"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc\") " Feb 25 13:46:24 crc kubenswrapper[4815]: I0225 13:46:24.856484 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-utilities" (OuterVolumeSpecName: "utilities") pod "f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc" (UID: "f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:46:24 crc kubenswrapper[4815]: I0225 13:46:24.874753 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-kube-api-access-sdxsz" (OuterVolumeSpecName: "kube-api-access-sdxsz") pod "f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc" (UID: "f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc"). InnerVolumeSpecName "kube-api-access-sdxsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:46:24 crc kubenswrapper[4815]: I0225 13:46:24.918909 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc" (UID: "f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:46:24 crc kubenswrapper[4815]: I0225 13:46:24.957730 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:46:24 crc kubenswrapper[4815]: I0225 13:46:24.958019 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdxsz\" (UniqueName: \"kubernetes.io/projected/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-kube-api-access-sdxsz\") on node \"crc\" DevicePath \"\"" Feb 25 13:46:24 crc kubenswrapper[4815]: I0225 13:46:24.958158 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:46:25 crc kubenswrapper[4815]: I0225 13:46:25.044979 4815 generic.go:334] "Generic (PLEG): container finished" podID="f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc" containerID="f94fd26d90a1ba7ce5481ebdce566b03cc4fa187919a2817bdad927c880a5288" exitCode=0 Feb 25 13:46:25 crc kubenswrapper[4815]: I0225 13:46:25.045992 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b4jcd" Feb 25 13:46:25 crc kubenswrapper[4815]: I0225 13:46:25.046422 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4jcd" event={"ID":"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc","Type":"ContainerDied","Data":"f94fd26d90a1ba7ce5481ebdce566b03cc4fa187919a2817bdad927c880a5288"} Feb 25 13:46:25 crc kubenswrapper[4815]: I0225 13:46:25.046448 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4jcd" event={"ID":"f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc","Type":"ContainerDied","Data":"4af03c7c52770acd061f8e3f24c878ae5e4063e17d7b90fe55849fa540ab16a9"} Feb 25 13:46:25 crc kubenswrapper[4815]: I0225 13:46:25.046464 4815 scope.go:117] "RemoveContainer" containerID="f94fd26d90a1ba7ce5481ebdce566b03cc4fa187919a2817bdad927c880a5288" Feb 25 13:46:25 crc kubenswrapper[4815]: I0225 13:46:25.071707 4815 scope.go:117] "RemoveContainer" containerID="9a5e6da638c04a2b0223119a8c05aafe33d9dcff08004fd86b264dc6add5a6fc" Feb 25 13:46:25 crc kubenswrapper[4815]: I0225 13:46:25.080737 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b4jcd"] Feb 25 13:46:25 crc kubenswrapper[4815]: I0225 13:46:25.089117 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b4jcd"] Feb 25 13:46:25 crc kubenswrapper[4815]: I0225 13:46:25.098061 4815 scope.go:117] "RemoveContainer" containerID="a4f0c05b7bfba0f138faed2f793181e370c1507683619e0a61e96199c88149f7" Feb 25 13:46:25 crc kubenswrapper[4815]: I0225 13:46:25.143228 4815 scope.go:117] "RemoveContainer" containerID="f94fd26d90a1ba7ce5481ebdce566b03cc4fa187919a2817bdad927c880a5288" Feb 25 13:46:25 crc kubenswrapper[4815]: E0225 13:46:25.144552 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f94fd26d90a1ba7ce5481ebdce566b03cc4fa187919a2817bdad927c880a5288\": container with ID starting with f94fd26d90a1ba7ce5481ebdce566b03cc4fa187919a2817bdad927c880a5288 not found: ID does not exist" containerID="f94fd26d90a1ba7ce5481ebdce566b03cc4fa187919a2817bdad927c880a5288" Feb 25 13:46:25 crc kubenswrapper[4815]: I0225 13:46:25.144602 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f94fd26d90a1ba7ce5481ebdce566b03cc4fa187919a2817bdad927c880a5288"} err="failed to get container status \"f94fd26d90a1ba7ce5481ebdce566b03cc4fa187919a2817bdad927c880a5288\": rpc error: code = NotFound desc = could not find container \"f94fd26d90a1ba7ce5481ebdce566b03cc4fa187919a2817bdad927c880a5288\": container with ID starting with f94fd26d90a1ba7ce5481ebdce566b03cc4fa187919a2817bdad927c880a5288 not found: ID does not exist" Feb 25 13:46:25 crc kubenswrapper[4815]: I0225 13:46:25.144631 4815 scope.go:117] "RemoveContainer" containerID="9a5e6da638c04a2b0223119a8c05aafe33d9dcff08004fd86b264dc6add5a6fc" Feb 25 13:46:25 crc kubenswrapper[4815]: E0225 13:46:25.144995 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a5e6da638c04a2b0223119a8c05aafe33d9dcff08004fd86b264dc6add5a6fc\": container with ID starting with 9a5e6da638c04a2b0223119a8c05aafe33d9dcff08004fd86b264dc6add5a6fc not found: ID does not exist" containerID="9a5e6da638c04a2b0223119a8c05aafe33d9dcff08004fd86b264dc6add5a6fc" Feb 25 13:46:25 crc kubenswrapper[4815]: I0225 13:46:25.145035 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a5e6da638c04a2b0223119a8c05aafe33d9dcff08004fd86b264dc6add5a6fc"} err="failed to get container status \"9a5e6da638c04a2b0223119a8c05aafe33d9dcff08004fd86b264dc6add5a6fc\": rpc error: code = NotFound desc = could not find container \"9a5e6da638c04a2b0223119a8c05aafe33d9dcff08004fd86b264dc6add5a6fc\": container with ID starting with 9a5e6da638c04a2b0223119a8c05aafe33d9dcff08004fd86b264dc6add5a6fc not found: ID does not exist" Feb 25 13:46:25 crc kubenswrapper[4815]: I0225 13:46:25.145062 4815 scope.go:117] "RemoveContainer" containerID="a4f0c05b7bfba0f138faed2f793181e370c1507683619e0a61e96199c88149f7" Feb 25 13:46:25 crc kubenswrapper[4815]: E0225 13:46:25.145331 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4f0c05b7bfba0f138faed2f793181e370c1507683619e0a61e96199c88149f7\": container with ID starting with a4f0c05b7bfba0f138faed2f793181e370c1507683619e0a61e96199c88149f7 not found: ID does not exist" containerID="a4f0c05b7bfba0f138faed2f793181e370c1507683619e0a61e96199c88149f7" Feb 25 13:46:25 crc kubenswrapper[4815]: I0225 13:46:25.145355 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4f0c05b7bfba0f138faed2f793181e370c1507683619e0a61e96199c88149f7"} err="failed to get container status \"a4f0c05b7bfba0f138faed2f793181e370c1507683619e0a61e96199c88149f7\": rpc error: code = NotFound desc = could not find container \"a4f0c05b7bfba0f138faed2f793181e370c1507683619e0a61e96199c88149f7\": container with ID starting with a4f0c05b7bfba0f138faed2f793181e370c1507683619e0a61e96199c88149f7 not found: ID does not exist" Feb 25 13:46:26 crc kubenswrapper[4815]: I0225 13:46:26.951681 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc" path="/var/lib/kubelet/pods/f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc/volumes" Feb 25 13:46:36 crc kubenswrapper[4815]: I0225 13:46:36.936169 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:46:36 crc kubenswrapper[4815]: E0225 13:46:36.937351 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:46:45 crc kubenswrapper[4815]: I0225 13:46:45.916219 4815 scope.go:117] "RemoveContainer" containerID="ac1b0e7571f7264de9e214a5d7804d337d6972277a4ca74433a716fbc1da94d9" Feb 25 13:46:45 crc kubenswrapper[4815]: I0225 13:46:45.942231 4815 scope.go:117] "RemoveContainer" containerID="18dd34d53e550f261ca45eb8f3d6ec974c09281859b2802e967d8144d1bfbda6" Feb 25 13:46:46 crc kubenswrapper[4815]: I0225 13:46:46.037358 4815 scope.go:117] "RemoveContainer" containerID="9c8e4523cc93cae849c9317f02d216c204cd816fb738b3b6ead4dd808da01f5e" Feb 25 13:46:48 crc kubenswrapper[4815]: I0225 13:46:48.936214 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:46:48 crc kubenswrapper[4815]: E0225 13:46:48.937189 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:46:56 crc kubenswrapper[4815]: I0225 13:46:56.052499 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-gp6dc"] Feb 25 13:46:56 crc kubenswrapper[4815]: I0225 13:46:56.068787 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-e541-account-create-update-7769t"] Feb 25 13:46:56 crc kubenswrapper[4815]: I0225 13:46:56.079568 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-4xctv"] Feb 25 13:46:56 crc kubenswrapper[4815]: I0225 13:46:56.089417 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-gp6dc"] Feb 25 13:46:56 crc kubenswrapper[4815]: I0225 13:46:56.098725 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-29cc-account-create-update-rf4f5"] Feb 25 13:46:56 crc kubenswrapper[4815]: I0225 13:46:56.110700 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-e541-account-create-update-7769t"] Feb 25 13:46:56 crc kubenswrapper[4815]: I0225 13:46:56.117215 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-4xctv"] Feb 25 13:46:56 crc kubenswrapper[4815]: I0225 13:46:56.124894 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-29cc-account-create-update-rf4f5"] Feb 25 13:46:56 crc kubenswrapper[4815]: I0225 13:46:56.947085 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50e8c23e-f13a-4fb9-b223-9330b3d1d36b" path="/var/lib/kubelet/pods/50e8c23e-f13a-4fb9-b223-9330b3d1d36b/volumes" Feb 25 13:46:56 crc kubenswrapper[4815]: I0225 13:46:56.948659 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64ed35f4-2af0-4d3e-84a7-6160ac5b9185" path="/var/lib/kubelet/pods/64ed35f4-2af0-4d3e-84a7-6160ac5b9185/volumes" Feb 25 13:46:56 crc kubenswrapper[4815]: I0225 13:46:56.949922 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84c1760e-09f1-4f63-ae74-8c8ba4af49be" path="/var/lib/kubelet/pods/84c1760e-09f1-4f63-ae74-8c8ba4af49be/volumes" Feb 25 13:46:56 crc kubenswrapper[4815]: I0225 13:46:56.950992 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a24929f6-66ff-415c-b539-a9e6ac507f1d" path="/var/lib/kubelet/pods/a24929f6-66ff-415c-b539-a9e6ac507f1d/volumes" Feb 25 13:46:59 crc kubenswrapper[4815]: I0225 13:46:59.041409 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-st8v2"] Feb 25 13:46:59 crc kubenswrapper[4815]: I0225 13:46:59.056278 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-st8v2"] Feb 25 13:47:00 crc kubenswrapper[4815]: I0225 13:47:00.029269 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-df70-account-create-update-s4qnr"] Feb 25 13:47:00 crc kubenswrapper[4815]: I0225 13:47:00.039374 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-df70-account-create-update-s4qnr"] Feb 25 13:47:00 crc kubenswrapper[4815]: I0225 13:47:00.945935 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="298a9862-55ab-4180-b10f-375379a96a2b" path="/var/lib/kubelet/pods/298a9862-55ab-4180-b10f-375379a96a2b/volumes" Feb 25 13:47:00 crc kubenswrapper[4815]: I0225 13:47:00.947048 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80276532-5b43-42b3-86db-de225388a3db" path="/var/lib/kubelet/pods/80276532-5b43-42b3-86db-de225388a3db/volumes" Feb 25 13:47:01 crc kubenswrapper[4815]: I0225 13:47:01.935912 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:47:01 crc kubenswrapper[4815]: E0225 13:47:01.936483 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:47:13 crc kubenswrapper[4815]: I0225 13:47:13.936982 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:47:13 crc kubenswrapper[4815]: E0225 13:47:13.937800 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:47:19 crc kubenswrapper[4815]: I0225 13:47:19.055024 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-49c8k"] Feb 25 13:47:19 crc kubenswrapper[4815]: I0225 13:47:19.065129 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-49c8k"] Feb 25 13:47:20 crc kubenswrapper[4815]: I0225 13:47:20.954245 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dbd3d76-08d8-4798-91ab-2dc5afe1993f" path="/var/lib/kubelet/pods/6dbd3d76-08d8-4798-91ab-2dc5afe1993f/volumes" Feb 25 13:47:25 crc kubenswrapper[4815]: I0225 13:47:25.041062 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-b9qrv"] Feb 25 13:47:25 crc kubenswrapper[4815]: I0225 13:47:25.048117 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-b9qrv"] Feb 25 13:47:25 crc kubenswrapper[4815]: I0225 13:47:25.935595 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:47:25 crc kubenswrapper[4815]: E0225 13:47:25.935840 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:47:26 crc kubenswrapper[4815]: I0225 13:47:26.956923 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e9844cb-1b51-4e42-8fee-5ed7059776c1" path="/var/lib/kubelet/pods/7e9844cb-1b51-4e42-8fee-5ed7059776c1/volumes" Feb 25 13:47:29 crc kubenswrapper[4815]: I0225 13:47:29.805447 4815 generic.go:334] "Generic (PLEG): container finished" podID="9d6402d2-d86e-4529-a4bb-b6b26f9ed63c" containerID="69873e4101a7a7278d6103f8c8fc828b6863b9922a48d96293639d50396b59a3" exitCode=0 Feb 25 13:47:29 crc kubenswrapper[4815]: I0225 13:47:29.805564 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" event={"ID":"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c","Type":"ContainerDied","Data":"69873e4101a7a7278d6103f8c8fc828b6863b9922a48d96293639d50396b59a3"} Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.262580 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.421793 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-ssh-key-openstack-edpm-ipam\") pod \"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c\" (UID: \"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c\") " Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.421889 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r4jw\" (UniqueName: \"kubernetes.io/projected/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-kube-api-access-5r4jw\") pod \"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c\" (UID: \"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c\") " Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.422146 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-inventory\") pod \"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c\" (UID: \"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c\") " Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.443564 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-kube-api-access-5r4jw" (OuterVolumeSpecName: "kube-api-access-5r4jw") pod "9d6402d2-d86e-4529-a4bb-b6b26f9ed63c" (UID: "9d6402d2-d86e-4529-a4bb-b6b26f9ed63c"). InnerVolumeSpecName "kube-api-access-5r4jw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.448002 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9d6402d2-d86e-4529-a4bb-b6b26f9ed63c" (UID: "9d6402d2-d86e-4529-a4bb-b6b26f9ed63c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.456053 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-inventory" (OuterVolumeSpecName: "inventory") pod "9d6402d2-d86e-4529-a4bb-b6b26f9ed63c" (UID: "9d6402d2-d86e-4529-a4bb-b6b26f9ed63c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.525014 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r4jw\" (UniqueName: \"kubernetes.io/projected/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-kube-api-access-5r4jw\") on node \"crc\" DevicePath \"\"" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.525072 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.525092 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9d6402d2-d86e-4529-a4bb-b6b26f9ed63c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.830056 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" event={"ID":"9d6402d2-d86e-4529-a4bb-b6b26f9ed63c","Type":"ContainerDied","Data":"c05dd6dc43365887ea0824b060077adc4109b377a4b77904e619dec4e6ea01de"} Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.830116 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c05dd6dc43365887ea0824b060077adc4109b377a4b77904e619dec4e6ea01de" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.830147 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.961006 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt"] Feb 25 13:47:31 crc kubenswrapper[4815]: E0225 13:47:31.961550 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc" containerName="registry-server" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.961570 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc" containerName="registry-server" Feb 25 13:47:31 crc kubenswrapper[4815]: E0225 13:47:31.961587 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc" containerName="extract-content" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.961598 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc" containerName="extract-content" Feb 25 13:47:31 crc kubenswrapper[4815]: E0225 13:47:31.961612 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc" containerName="extract-utilities" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.961620 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc" containerName="extract-utilities" Feb 25 13:47:31 crc kubenswrapper[4815]: E0225 13:47:31.961632 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d6402d2-d86e-4529-a4bb-b6b26f9ed63c" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.961641 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d6402d2-d86e-4529-a4bb-b6b26f9ed63c" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.961871 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d6402d2-d86e-4529-a4bb-b6b26f9ed63c" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.961895 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b25aff-3bb2-4fa7-9eea-ce5fc06b77cc" containerName="registry-server" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.962682 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.965020 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.965738 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.965936 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mmbj" Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.971986 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt"] Feb 25 13:47:31 crc kubenswrapper[4815]: I0225 13:47:31.973646 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 25 13:47:32 crc kubenswrapper[4815]: I0225 13:47:32.035157 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52d4b958-a8c5-42ee-bd46-d541d47629cb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt\" (UID: \"52d4b958-a8c5-42ee-bd46-d541d47629cb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" Feb 25 13:47:32 crc kubenswrapper[4815]: I0225 13:47:32.035224 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrr6g\" (UniqueName: \"kubernetes.io/projected/52d4b958-a8c5-42ee-bd46-d541d47629cb-kube-api-access-nrr6g\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt\" (UID: \"52d4b958-a8c5-42ee-bd46-d541d47629cb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" Feb 25 13:47:32 crc kubenswrapper[4815]: I0225 13:47:32.037001 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/52d4b958-a8c5-42ee-bd46-d541d47629cb-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt\" (UID: \"52d4b958-a8c5-42ee-bd46-d541d47629cb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" Feb 25 13:47:32 crc kubenswrapper[4815]: I0225 13:47:32.139109 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/52d4b958-a8c5-42ee-bd46-d541d47629cb-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt\" (UID: \"52d4b958-a8c5-42ee-bd46-d541d47629cb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" Feb 25 13:47:32 crc kubenswrapper[4815]: I0225 13:47:32.139731 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52d4b958-a8c5-42ee-bd46-d541d47629cb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt\" (UID: \"52d4b958-a8c5-42ee-bd46-d541d47629cb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" Feb 25 13:47:32 crc kubenswrapper[4815]: I0225 13:47:32.139796 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrr6g\" (UniqueName: \"kubernetes.io/projected/52d4b958-a8c5-42ee-bd46-d541d47629cb-kube-api-access-nrr6g\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt\" (UID: \"52d4b958-a8c5-42ee-bd46-d541d47629cb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" Feb 25 13:47:32 crc kubenswrapper[4815]: I0225 13:47:32.143360 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52d4b958-a8c5-42ee-bd46-d541d47629cb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt\" (UID: \"52d4b958-a8c5-42ee-bd46-d541d47629cb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" Feb 25 13:47:32 crc kubenswrapper[4815]: I0225 13:47:32.143600 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/52d4b958-a8c5-42ee-bd46-d541d47629cb-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt\" (UID: \"52d4b958-a8c5-42ee-bd46-d541d47629cb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" Feb 25 13:47:32 crc kubenswrapper[4815]: I0225 13:47:32.164299 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrr6g\" (UniqueName: \"kubernetes.io/projected/52d4b958-a8c5-42ee-bd46-d541d47629cb-kube-api-access-nrr6g\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt\" (UID: \"52d4b958-a8c5-42ee-bd46-d541d47629cb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" Feb 25 13:47:32 crc kubenswrapper[4815]: I0225 13:47:32.287376 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" Feb 25 13:47:32 crc kubenswrapper[4815]: I0225 13:47:32.660363 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt"] Feb 25 13:47:32 crc kubenswrapper[4815]: I0225 13:47:32.839991 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" event={"ID":"52d4b958-a8c5-42ee-bd46-d541d47629cb","Type":"ContainerStarted","Data":"f578ee78a40c69ccc260d97cd019d160efc850105593eb7380d8ffc6963387d1"} Feb 25 13:47:33 crc kubenswrapper[4815]: I0225 13:47:33.031225 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-bc7lz"] Feb 25 13:47:33 crc kubenswrapper[4815]: I0225 13:47:33.042994 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-bc7lz"] Feb 25 13:47:33 crc kubenswrapper[4815]: I0225 13:47:33.852175 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" event={"ID":"52d4b958-a8c5-42ee-bd46-d541d47629cb","Type":"ContainerStarted","Data":"afdd9855d7996e75bb1cdde0b01d4a3dce75b35201834ddfea4a141781659666"} Feb 25 13:47:33 crc kubenswrapper[4815]: I0225 13:47:33.876954 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" podStartSLOduration=2.361840526 podStartE2EDuration="2.876937016s" podCreationTimestamp="2026-02-25 13:47:31 +0000 UTC" firstStartedPulling="2026-02-25 13:47:32.65931828 +0000 UTC m=+1610.460416334" lastFinishedPulling="2026-02-25 13:47:33.17441477 +0000 UTC m=+1610.975512824" observedRunningTime="2026-02-25 13:47:33.872045935 +0000 UTC m=+1611.673144019" watchObservedRunningTime="2026-02-25 13:47:33.876937016 +0000 UTC m=+1611.678035070" Feb 25 13:47:34 crc kubenswrapper[4815]: I0225 13:47:34.048191 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-55a3-account-create-update-vjcsr"] Feb 25 13:47:34 crc kubenswrapper[4815]: I0225 13:47:34.060238 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-55a3-account-create-update-vjcsr"] Feb 25 13:47:34 crc kubenswrapper[4815]: I0225 13:47:34.951602 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80bad2a7-eed6-4d5e-a631-160681e10e41" path="/var/lib/kubelet/pods/80bad2a7-eed6-4d5e-a631-160681e10e41/volumes" Feb 25 13:47:34 crc kubenswrapper[4815]: I0225 13:47:34.952901 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7b99305-c70e-422b-925c-d9abfc038a5e" path="/var/lib/kubelet/pods/d7b99305-c70e-422b-925c-d9abfc038a5e/volumes" Feb 25 13:47:35 crc kubenswrapper[4815]: I0225 13:47:35.060757 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-6hvck"] Feb 25 13:47:35 crc kubenswrapper[4815]: I0225 13:47:35.065974 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-198e-account-create-update-t9r4s"] Feb 25 13:47:35 crc kubenswrapper[4815]: I0225 13:47:35.076315 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-ce70-account-create-update-75n7s"] Feb 25 13:47:35 crc kubenswrapper[4815]: I0225 13:47:35.085478 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-lrkj8"] Feb 25 13:47:35 crc kubenswrapper[4815]: I0225 13:47:35.094577 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-6hvck"] Feb 25 13:47:35 crc kubenswrapper[4815]: I0225 13:47:35.103885 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-lrkj8"] Feb 25 13:47:35 crc kubenswrapper[4815]: I0225 13:47:35.113471 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-198e-account-create-update-t9r4s"] Feb 25 13:47:35 crc kubenswrapper[4815]: I0225 13:47:35.122398 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-ce70-account-create-update-75n7s"] Feb 25 13:47:36 crc kubenswrapper[4815]: I0225 13:47:36.936866 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:47:36 crc kubenswrapper[4815]: E0225 13:47:36.937650 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:47:36 crc kubenswrapper[4815]: I0225 13:47:36.945205 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1095ac90-8cd8-41d8-8eb7-847f72d3d82b" path="/var/lib/kubelet/pods/1095ac90-8cd8-41d8-8eb7-847f72d3d82b/volumes" Feb 25 13:47:36 crc kubenswrapper[4815]: I0225 13:47:36.945892 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e394112-2dbe-43b0-8d71-5763831f591a" path="/var/lib/kubelet/pods/2e394112-2dbe-43b0-8d71-5763831f591a/volumes" Feb 25 13:47:36 crc kubenswrapper[4815]: I0225 13:47:36.946633 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a45cad8-f93f-4bf1-9fc8-d382e6be15f7" path="/var/lib/kubelet/pods/7a45cad8-f93f-4bf1-9fc8-d382e6be15f7/volumes" Feb 25 13:47:36 crc kubenswrapper[4815]: I0225 13:47:36.947272 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b912f0ba-4a07-4b2a-b372-a107335c4b73" path="/var/lib/kubelet/pods/b912f0ba-4a07-4b2a-b372-a107335c4b73/volumes" Feb 25 13:47:43 crc kubenswrapper[4815]: I0225 13:47:43.043190 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-szm5v"] Feb 25 13:47:43 crc kubenswrapper[4815]: I0225 13:47:43.051267 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-szm5v"] Feb 25 13:47:44 crc kubenswrapper[4815]: I0225 13:47:44.956453 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75268cc3-204e-4ac5-9cbd-958343dd2011" path="/var/lib/kubelet/pods/75268cc3-204e-4ac5-9cbd-958343dd2011/volumes" Feb 25 13:47:46 crc kubenswrapper[4815]: I0225 13:47:46.166875 4815 scope.go:117] "RemoveContainer" containerID="9e48c2504573e6f135e3796d7ae6705ac4f3d7b6a34fe030c1ba321ee632844a" Feb 25 13:47:46 crc kubenswrapper[4815]: I0225 13:47:46.221815 4815 scope.go:117] "RemoveContainer" containerID="96f99d91e60be382a1ff85ca50faf71803c3409dd984d5cb89b984c01d94cf00" Feb 25 13:47:46 crc kubenswrapper[4815]: I0225 13:47:46.257048 4815 scope.go:117] "RemoveContainer" containerID="32bb1bba6e9f9bc6ce15dd3c3911076bc6a8ef51fc9788763a3d5e23b1ef529b" Feb 25 13:47:46 crc kubenswrapper[4815]: I0225 13:47:46.298244 4815 scope.go:117] "RemoveContainer" containerID="16f6265b307311743431a947d38dd872750903a01059624c86677b8b18aa1581" Feb 25 13:47:46 crc kubenswrapper[4815]: I0225 13:47:46.339386 4815 scope.go:117] "RemoveContainer" containerID="4fe814b94077353d6a8d2f48e8ce07000a7d18da25c3e7efec37bce1ec6f0c06" Feb 25 13:47:46 crc kubenswrapper[4815]: I0225 13:47:46.393481 4815 scope.go:117] "RemoveContainer" containerID="c8098a5c9f4fded8b1aa8dc1f4360ba63f0c8505eb27faffe83a00e7c0741271" Feb 25 13:47:46 crc kubenswrapper[4815]: I0225 13:47:46.426498 4815 scope.go:117] "RemoveContainer" containerID="0795609e27a96f4062f243e6c5401f60a1d133919c6b727cfab7be2b687b257b" Feb 25 13:47:46 crc kubenswrapper[4815]: I0225 13:47:46.446910 4815 scope.go:117] "RemoveContainer" containerID="a5c9d925676cb371db86aa6cc24221d97964582b03c81997911af246f539ca24" Feb 25 13:47:46 crc kubenswrapper[4815]: I0225 13:47:46.467367 4815 scope.go:117] "RemoveContainer" containerID="e7700b21016181b8e1d769c3aee588c4178c7221decd73961bca470edef59d89" Feb 25 13:47:46 crc kubenswrapper[4815]: I0225 13:47:46.490846 4815 scope.go:117] "RemoveContainer" containerID="f2018e423eff7617abcb82018b6311c51dd8916a5d122045de02b6cc8e8e7684" Feb 25 13:47:46 crc kubenswrapper[4815]: I0225 13:47:46.529904 4815 scope.go:117] "RemoveContainer" containerID="01e045a2656f40d52c581376e503f61bb812e34ebe0a3607a28c08beaf4adedf" Feb 25 13:47:46 crc kubenswrapper[4815]: I0225 13:47:46.555077 4815 scope.go:117] "RemoveContainer" containerID="a6084341b178e86ba2f4e30ef013ac0618b5b575ca28c75ef03b15cd6a863aaa" Feb 25 13:47:46 crc kubenswrapper[4815]: I0225 13:47:46.585482 4815 scope.go:117] "RemoveContainer" containerID="920719f20c53bbaee0c90eb85e6b003ac665d58ab7ced9d807566981c86c04e7" Feb 25 13:47:46 crc kubenswrapper[4815]: I0225 13:47:46.618164 4815 scope.go:117] "RemoveContainer" containerID="af0e9370809232daebcb9ce84280e56c4c962c934ca4372cff8f06aadf0c8259" Feb 25 13:47:46 crc kubenswrapper[4815]: I0225 13:47:46.638309 4815 scope.go:117] "RemoveContainer" containerID="412bba9239a0b0e8d3a6b7639584f77e2da023f720ddb4cad3e9fbbde36761b0" Feb 25 13:47:51 crc kubenswrapper[4815]: I0225 13:47:51.936636 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:47:51 crc kubenswrapper[4815]: E0225 13:47:51.937416 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:48:00 crc kubenswrapper[4815]: I0225 13:48:00.167006 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533788-7kz4t"] Feb 25 13:48:00 crc kubenswrapper[4815]: I0225 13:48:00.169612 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533788-7kz4t" Feb 25 13:48:00 crc kubenswrapper[4815]: I0225 13:48:00.173563 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:48:00 crc kubenswrapper[4815]: I0225 13:48:00.173808 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:48:00 crc kubenswrapper[4815]: I0225 13:48:00.177140 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:48:00 crc kubenswrapper[4815]: I0225 13:48:00.183685 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533788-7kz4t"] Feb 25 13:48:00 crc kubenswrapper[4815]: I0225 13:48:00.344165 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg4mr\" (UniqueName: \"kubernetes.io/projected/9533594f-c28a-46a8-b87b-2d872b53d1b3-kube-api-access-cg4mr\") pod \"auto-csr-approver-29533788-7kz4t\" (UID: \"9533594f-c28a-46a8-b87b-2d872b53d1b3\") " pod="openshift-infra/auto-csr-approver-29533788-7kz4t" Feb 25 13:48:00 crc kubenswrapper[4815]: I0225 13:48:00.446609 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg4mr\" (UniqueName: \"kubernetes.io/projected/9533594f-c28a-46a8-b87b-2d872b53d1b3-kube-api-access-cg4mr\") pod \"auto-csr-approver-29533788-7kz4t\" (UID: \"9533594f-c28a-46a8-b87b-2d872b53d1b3\") " pod="openshift-infra/auto-csr-approver-29533788-7kz4t" Feb 25 13:48:00 crc kubenswrapper[4815]: I0225 13:48:00.472356 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg4mr\" (UniqueName: \"kubernetes.io/projected/9533594f-c28a-46a8-b87b-2d872b53d1b3-kube-api-access-cg4mr\") pod \"auto-csr-approver-29533788-7kz4t\" (UID: \"9533594f-c28a-46a8-b87b-2d872b53d1b3\") " pod="openshift-infra/auto-csr-approver-29533788-7kz4t" Feb 25 13:48:00 crc kubenswrapper[4815]: I0225 13:48:00.504637 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533788-7kz4t" Feb 25 13:48:00 crc kubenswrapper[4815]: I0225 13:48:00.991276 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533788-7kz4t"] Feb 25 13:48:01 crc kubenswrapper[4815]: I0225 13:48:01.161799 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533788-7kz4t" event={"ID":"9533594f-c28a-46a8-b87b-2d872b53d1b3","Type":"ContainerStarted","Data":"b11de6c86a664b8e49c8dbb2b9da93a8e42a46f2b1919646c1d1dd48f386c448"} Feb 25 13:48:03 crc kubenswrapper[4815]: I0225 13:48:03.180296 4815 generic.go:334] "Generic (PLEG): container finished" podID="9533594f-c28a-46a8-b87b-2d872b53d1b3" containerID="827676a0a42a6d72a7625fbd961d1c9d4cba3ce0e673f3a31d2f446c05118cd9" exitCode=0 Feb 25 13:48:03 crc kubenswrapper[4815]: I0225 13:48:03.180344 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533788-7kz4t" event={"ID":"9533594f-c28a-46a8-b87b-2d872b53d1b3","Type":"ContainerDied","Data":"827676a0a42a6d72a7625fbd961d1c9d4cba3ce0e673f3a31d2f446c05118cd9"} Feb 25 13:48:04 crc kubenswrapper[4815]: I0225 13:48:04.541448 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533788-7kz4t" Feb 25 13:48:04 crc kubenswrapper[4815]: I0225 13:48:04.740982 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cg4mr\" (UniqueName: \"kubernetes.io/projected/9533594f-c28a-46a8-b87b-2d872b53d1b3-kube-api-access-cg4mr\") pod \"9533594f-c28a-46a8-b87b-2d872b53d1b3\" (UID: \"9533594f-c28a-46a8-b87b-2d872b53d1b3\") " Feb 25 13:48:04 crc kubenswrapper[4815]: I0225 13:48:04.748256 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9533594f-c28a-46a8-b87b-2d872b53d1b3-kube-api-access-cg4mr" (OuterVolumeSpecName: "kube-api-access-cg4mr") pod "9533594f-c28a-46a8-b87b-2d872b53d1b3" (UID: "9533594f-c28a-46a8-b87b-2d872b53d1b3"). InnerVolumeSpecName "kube-api-access-cg4mr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:48:04 crc kubenswrapper[4815]: I0225 13:48:04.843746 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cg4mr\" (UniqueName: \"kubernetes.io/projected/9533594f-c28a-46a8-b87b-2d872b53d1b3-kube-api-access-cg4mr\") on node \"crc\" DevicePath \"\"" Feb 25 13:48:05 crc kubenswrapper[4815]: I0225 13:48:05.207163 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533788-7kz4t" event={"ID":"9533594f-c28a-46a8-b87b-2d872b53d1b3","Type":"ContainerDied","Data":"b11de6c86a664b8e49c8dbb2b9da93a8e42a46f2b1919646c1d1dd48f386c448"} Feb 25 13:48:05 crc kubenswrapper[4815]: I0225 13:48:05.207238 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b11de6c86a664b8e49c8dbb2b9da93a8e42a46f2b1919646c1d1dd48f386c448" Feb 25 13:48:05 crc kubenswrapper[4815]: I0225 13:48:05.207267 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533788-7kz4t" Feb 25 13:48:05 crc kubenswrapper[4815]: I0225 13:48:05.639842 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533782-cw687"] Feb 25 13:48:05 crc kubenswrapper[4815]: I0225 13:48:05.653811 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533782-cw687"] Feb 25 13:48:06 crc kubenswrapper[4815]: I0225 13:48:06.936343 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:48:06 crc kubenswrapper[4815]: E0225 13:48:06.936816 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:48:06 crc kubenswrapper[4815]: I0225 13:48:06.948933 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34838863-42b3-440d-96be-a1a0c48b32ed" path="/var/lib/kubelet/pods/34838863-42b3-440d-96be-a1a0c48b32ed/volumes" Feb 25 13:48:12 crc kubenswrapper[4815]: I0225 13:48:12.034536 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-2kr57"] Feb 25 13:48:12 crc kubenswrapper[4815]: I0225 13:48:12.044475 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-2kr57"] Feb 25 13:48:12 crc kubenswrapper[4815]: I0225 13:48:12.957891 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab03b49b-be69-4e1d-99e7-2f71a602b110" path="/var/lib/kubelet/pods/ab03b49b-be69-4e1d-99e7-2f71a602b110/volumes" Feb 25 13:48:19 crc kubenswrapper[4815]: I0225 13:48:19.045460 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-mgztm"] Feb 25 13:48:19 crc kubenswrapper[4815]: I0225 13:48:19.069871 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-mgztm"] Feb 25 13:48:20 crc kubenswrapper[4815]: I0225 13:48:20.950585 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee89a5e7-e5a4-412f-a52a-511dfca61265" path="/var/lib/kubelet/pods/ee89a5e7-e5a4-412f-a52a-511dfca61265/volumes" Feb 25 13:48:21 crc kubenswrapper[4815]: I0225 13:48:21.936424 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:48:21 crc kubenswrapper[4815]: E0225 13:48:21.936709 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:48:22 crc kubenswrapper[4815]: I0225 13:48:22.037071 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-4x94p"] Feb 25 13:48:22 crc kubenswrapper[4815]: I0225 13:48:22.046941 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-4x94p"] Feb 25 13:48:22 crc kubenswrapper[4815]: I0225 13:48:22.967855 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63427264-5351-40c4-aeff-d35ea829ccd4" path="/var/lib/kubelet/pods/63427264-5351-40c4-aeff-d35ea829ccd4/volumes" Feb 25 13:48:36 crc kubenswrapper[4815]: I0225 13:48:36.581357 4815 generic.go:334] "Generic (PLEG): container finished" podID="52d4b958-a8c5-42ee-bd46-d541d47629cb" containerID="afdd9855d7996e75bb1cdde0b01d4a3dce75b35201834ddfea4a141781659666" exitCode=0 Feb 25 13:48:36 crc kubenswrapper[4815]: I0225 13:48:36.581500 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" event={"ID":"52d4b958-a8c5-42ee-bd46-d541d47629cb","Type":"ContainerDied","Data":"afdd9855d7996e75bb1cdde0b01d4a3dce75b35201834ddfea4a141781659666"} Feb 25 13:48:36 crc kubenswrapper[4815]: I0225 13:48:36.936117 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:48:36 crc kubenswrapper[4815]: E0225 13:48:36.936527 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:48:37 crc kubenswrapper[4815]: I0225 13:48:37.046761 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-d4ssg"] Feb 25 13:48:37 crc kubenswrapper[4815]: I0225 13:48:37.055943 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-d4ssg"] Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.030225 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.119222 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/52d4b958-a8c5-42ee-bd46-d541d47629cb-ssh-key-openstack-edpm-ipam\") pod \"52d4b958-a8c5-42ee-bd46-d541d47629cb\" (UID: \"52d4b958-a8c5-42ee-bd46-d541d47629cb\") " Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.119550 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrr6g\" (UniqueName: \"kubernetes.io/projected/52d4b958-a8c5-42ee-bd46-d541d47629cb-kube-api-access-nrr6g\") pod \"52d4b958-a8c5-42ee-bd46-d541d47629cb\" (UID: \"52d4b958-a8c5-42ee-bd46-d541d47629cb\") " Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.119699 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52d4b958-a8c5-42ee-bd46-d541d47629cb-inventory\") pod \"52d4b958-a8c5-42ee-bd46-d541d47629cb\" (UID: \"52d4b958-a8c5-42ee-bd46-d541d47629cb\") " Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.127235 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52d4b958-a8c5-42ee-bd46-d541d47629cb-kube-api-access-nrr6g" (OuterVolumeSpecName: "kube-api-access-nrr6g") pod "52d4b958-a8c5-42ee-bd46-d541d47629cb" (UID: "52d4b958-a8c5-42ee-bd46-d541d47629cb"). InnerVolumeSpecName "kube-api-access-nrr6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.153970 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52d4b958-a8c5-42ee-bd46-d541d47629cb-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "52d4b958-a8c5-42ee-bd46-d541d47629cb" (UID: "52d4b958-a8c5-42ee-bd46-d541d47629cb"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.174660 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52d4b958-a8c5-42ee-bd46-d541d47629cb-inventory" (OuterVolumeSpecName: "inventory") pod "52d4b958-a8c5-42ee-bd46-d541d47629cb" (UID: "52d4b958-a8c5-42ee-bd46-d541d47629cb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.222665 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/52d4b958-a8c5-42ee-bd46-d541d47629cb-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.222709 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrr6g\" (UniqueName: \"kubernetes.io/projected/52d4b958-a8c5-42ee-bd46-d541d47629cb-kube-api-access-nrr6g\") on node \"crc\" DevicePath \"\"" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.222719 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52d4b958-a8c5-42ee-bd46-d541d47629cb-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.605311 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" event={"ID":"52d4b958-a8c5-42ee-bd46-d541d47629cb","Type":"ContainerDied","Data":"f578ee78a40c69ccc260d97cd019d160efc850105593eb7380d8ffc6963387d1"} Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.605610 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f578ee78a40c69ccc260d97cd019d160efc850105593eb7380d8ffc6963387d1" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.605397 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.707101 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb"] Feb 25 13:48:38 crc kubenswrapper[4815]: E0225 13:48:38.707812 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52d4b958-a8c5-42ee-bd46-d541d47629cb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.709168 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="52d4b958-a8c5-42ee-bd46-d541d47629cb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 25 13:48:38 crc kubenswrapper[4815]: E0225 13:48:38.709315 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9533594f-c28a-46a8-b87b-2d872b53d1b3" containerName="oc" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.709380 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9533594f-c28a-46a8-b87b-2d872b53d1b3" containerName="oc" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.709700 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="52d4b958-a8c5-42ee-bd46-d541d47629cb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.709789 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="9533594f-c28a-46a8-b87b-2d872b53d1b3" containerName="oc" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.710627 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.715800 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mmbj" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.716152 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.716397 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.716564 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.719492 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb"] Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.834387 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm5l5\" (UniqueName: \"kubernetes.io/projected/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-kube-api-access-nm5l5\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb\" (UID: \"9fdfaae5-e949-48ed-b3cf-a8893dc84a67\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.834526 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb\" (UID: \"9fdfaae5-e949-48ed-b3cf-a8893dc84a67\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.834597 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb\" (UID: \"9fdfaae5-e949-48ed-b3cf-a8893dc84a67\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.936284 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb\" (UID: \"9fdfaae5-e949-48ed-b3cf-a8893dc84a67\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.936393 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm5l5\" (UniqueName: \"kubernetes.io/projected/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-kube-api-access-nm5l5\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb\" (UID: \"9fdfaae5-e949-48ed-b3cf-a8893dc84a67\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.936475 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb\" (UID: \"9fdfaae5-e949-48ed-b3cf-a8893dc84a67\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.940791 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb\" (UID: \"9fdfaae5-e949-48ed-b3cf-a8893dc84a67\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.940802 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb\" (UID: \"9fdfaae5-e949-48ed-b3cf-a8893dc84a67\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.973626 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm5l5\" (UniqueName: \"kubernetes.io/projected/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-kube-api-access-nm5l5\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb\" (UID: \"9fdfaae5-e949-48ed-b3cf-a8893dc84a67\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" Feb 25 13:48:38 crc kubenswrapper[4815]: I0225 13:48:38.975042 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bdc9071-e863-4108-9234-4f201732f435" path="/var/lib/kubelet/pods/7bdc9071-e863-4108-9234-4f201732f435/volumes" Feb 25 13:48:39 crc kubenswrapper[4815]: I0225 13:48:39.034828 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" Feb 25 13:48:39 crc kubenswrapper[4815]: I0225 13:48:39.047977 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-576gs"] Feb 25 13:48:39 crc kubenswrapper[4815]: I0225 13:48:39.061621 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-576gs"] Feb 25 13:48:39 crc kubenswrapper[4815]: I0225 13:48:39.584649 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb"] Feb 25 13:48:39 crc kubenswrapper[4815]: I0225 13:48:39.615658 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" event={"ID":"9fdfaae5-e949-48ed-b3cf-a8893dc84a67","Type":"ContainerStarted","Data":"036855b9a9610399b6b48f4b9f0dfc7cd02b5d9298e3d4d0675863fba9abdb17"} Feb 25 13:48:40 crc kubenswrapper[4815]: I0225 13:48:40.625623 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" event={"ID":"9fdfaae5-e949-48ed-b3cf-a8893dc84a67","Type":"ContainerStarted","Data":"340723b43f42e90b91a4805b3dd5ac42edb2446637fc609ee789dfa3c89020c5"} Feb 25 13:48:40 crc kubenswrapper[4815]: I0225 13:48:40.672175 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" podStartSLOduration=2.11958917 podStartE2EDuration="2.672148411s" podCreationTimestamp="2026-02-25 13:48:38 +0000 UTC" firstStartedPulling="2026-02-25 13:48:39.593426972 +0000 UTC m=+1677.394525036" lastFinishedPulling="2026-02-25 13:48:40.145986223 +0000 UTC m=+1677.947084277" observedRunningTime="2026-02-25 13:48:40.645102377 +0000 UTC m=+1678.446200461" watchObservedRunningTime="2026-02-25 13:48:40.672148411 +0000 UTC m=+1678.473246465" Feb 25 13:48:40 crc kubenswrapper[4815]: I0225 13:48:40.950623 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04e3d721-2179-4457-8cc6-72af3b3f9257" path="/var/lib/kubelet/pods/04e3d721-2179-4457-8cc6-72af3b3f9257/volumes" Feb 25 13:48:44 crc kubenswrapper[4815]: E0225 13:48:44.929864 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fdfaae5_e949_48ed_b3cf_a8893dc84a67.slice/crio-conmon-340723b43f42e90b91a4805b3dd5ac42edb2446637fc609ee789dfa3c89020c5.scope\": RecentStats: unable to find data in memory cache]" Feb 25 13:48:45 crc kubenswrapper[4815]: I0225 13:48:45.682210 4815 generic.go:334] "Generic (PLEG): container finished" podID="9fdfaae5-e949-48ed-b3cf-a8893dc84a67" containerID="340723b43f42e90b91a4805b3dd5ac42edb2446637fc609ee789dfa3c89020c5" exitCode=0 Feb 25 13:48:45 crc kubenswrapper[4815]: I0225 13:48:45.682253 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" event={"ID":"9fdfaae5-e949-48ed-b3cf-a8893dc84a67","Type":"ContainerDied","Data":"340723b43f42e90b91a4805b3dd5ac42edb2446637fc609ee789dfa3c89020c5"} Feb 25 13:48:46 crc kubenswrapper[4815]: I0225 13:48:46.956748 4815 scope.go:117] "RemoveContainer" containerID="4dd4c4d4dba488e7472dcc7279f6c3ad41d76de7b59eb5521533e41a8e0c8a40" Feb 25 13:48:46 crc kubenswrapper[4815]: I0225 13:48:46.996875 4815 scope.go:117] "RemoveContainer" containerID="c5a9af64e2d4e57b092b43e4a47fe7a5daffc4c0026054c3e700d2da1bb3823d" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.036080 4815 scope.go:117] "RemoveContainer" containerID="ebb6d812cafbfdf225df72faccc76a98ea8fc505a950399f7bcacf5169f65982" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.141105 4815 scope.go:117] "RemoveContainer" containerID="aa83b80cdebfae2c0db1d74c7475c986c41dd26e06d53c3e44db3c6c664fa6df" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.162887 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.202306 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-ssh-key-openstack-edpm-ipam\") pod \"9fdfaae5-e949-48ed-b3cf-a8893dc84a67\" (UID: \"9fdfaae5-e949-48ed-b3cf-a8893dc84a67\") " Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.202410 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-inventory\") pod \"9fdfaae5-e949-48ed-b3cf-a8893dc84a67\" (UID: \"9fdfaae5-e949-48ed-b3cf-a8893dc84a67\") " Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.202455 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nm5l5\" (UniqueName: \"kubernetes.io/projected/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-kube-api-access-nm5l5\") pod \"9fdfaae5-e949-48ed-b3cf-a8893dc84a67\" (UID: \"9fdfaae5-e949-48ed-b3cf-a8893dc84a67\") " Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.206988 4815 scope.go:117] "RemoveContainer" containerID="4ce2c3c9b64d5f047c35dc1e7ac70a87bf7be1bc271d659eef2e7ae14c2e1a8e" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.209080 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-kube-api-access-nm5l5" (OuterVolumeSpecName: "kube-api-access-nm5l5") pod "9fdfaae5-e949-48ed-b3cf-a8893dc84a67" (UID: "9fdfaae5-e949-48ed-b3cf-a8893dc84a67"). InnerVolumeSpecName "kube-api-access-nm5l5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.240205 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9fdfaae5-e949-48ed-b3cf-a8893dc84a67" (UID: "9fdfaae5-e949-48ed-b3cf-a8893dc84a67"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.247803 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-inventory" (OuterVolumeSpecName: "inventory") pod "9fdfaae5-e949-48ed-b3cf-a8893dc84a67" (UID: "9fdfaae5-e949-48ed-b3cf-a8893dc84a67"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.255626 4815 scope.go:117] "RemoveContainer" containerID="9d5688a7d91b39b94743229dfb65ac8852aa764ec59864066e8ded936e9d1ece" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.304762 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.304800 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.304814 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nm5l5\" (UniqueName: \"kubernetes.io/projected/9fdfaae5-e949-48ed-b3cf-a8893dc84a67-kube-api-access-nm5l5\") on node \"crc\" DevicePath \"\"" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.703378 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" event={"ID":"9fdfaae5-e949-48ed-b3cf-a8893dc84a67","Type":"ContainerDied","Data":"036855b9a9610399b6b48f4b9f0dfc7cd02b5d9298e3d4d0675863fba9abdb17"} Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.703416 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="036855b9a9610399b6b48f4b9f0dfc7cd02b5d9298e3d4d0675863fba9abdb17" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.703476 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.793153 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs"] Feb 25 13:48:47 crc kubenswrapper[4815]: E0225 13:48:47.793609 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fdfaae5-e949-48ed-b3cf-a8893dc84a67" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.793630 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fdfaae5-e949-48ed-b3cf-a8893dc84a67" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.793837 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fdfaae5-e949-48ed-b3cf-a8893dc84a67" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.794634 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.797084 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mmbj" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.797552 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.797758 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.803034 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.812285 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6add50b-b148-4e9b-af2a-36571841b3c0-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwbs\" (UID: \"d6add50b-b148-4e9b-af2a-36571841b3c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.812334 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9vxd\" (UniqueName: \"kubernetes.io/projected/d6add50b-b148-4e9b-af2a-36571841b3c0-kube-api-access-d9vxd\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwbs\" (UID: \"d6add50b-b148-4e9b-af2a-36571841b3c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.812410 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d6add50b-b148-4e9b-af2a-36571841b3c0-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwbs\" (UID: \"d6add50b-b148-4e9b-af2a-36571841b3c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.817468 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs"] Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.913092 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6add50b-b148-4e9b-af2a-36571841b3c0-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwbs\" (UID: \"d6add50b-b148-4e9b-af2a-36571841b3c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.913143 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9vxd\" (UniqueName: \"kubernetes.io/projected/d6add50b-b148-4e9b-af2a-36571841b3c0-kube-api-access-d9vxd\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwbs\" (UID: \"d6add50b-b148-4e9b-af2a-36571841b3c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.913211 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d6add50b-b148-4e9b-af2a-36571841b3c0-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwbs\" (UID: \"d6add50b-b148-4e9b-af2a-36571841b3c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.916735 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d6add50b-b148-4e9b-af2a-36571841b3c0-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwbs\" (UID: \"d6add50b-b148-4e9b-af2a-36571841b3c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.916744 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6add50b-b148-4e9b-af2a-36571841b3c0-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwbs\" (UID: \"d6add50b-b148-4e9b-af2a-36571841b3c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" Feb 25 13:48:47 crc kubenswrapper[4815]: I0225 13:48:47.934273 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9vxd\" (UniqueName: \"kubernetes.io/projected/d6add50b-b148-4e9b-af2a-36571841b3c0-kube-api-access-d9vxd\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sgwbs\" (UID: \"d6add50b-b148-4e9b-af2a-36571841b3c0\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" Feb 25 13:48:48 crc kubenswrapper[4815]: I0225 13:48:48.114396 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" Feb 25 13:48:48 crc kubenswrapper[4815]: I0225 13:48:48.655930 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs"] Feb 25 13:48:48 crc kubenswrapper[4815]: W0225 13:48:48.658306 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6add50b_b148_4e9b_af2a_36571841b3c0.slice/crio-99b849db1e9d702ce17759fc56ec702756241a9a61ca38bdef68a1003f351016 WatchSource:0}: Error finding container 99b849db1e9d702ce17759fc56ec702756241a9a61ca38bdef68a1003f351016: Status 404 returned error can't find the container with id 99b849db1e9d702ce17759fc56ec702756241a9a61ca38bdef68a1003f351016 Feb 25 13:48:48 crc kubenswrapper[4815]: I0225 13:48:48.712721 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" event={"ID":"d6add50b-b148-4e9b-af2a-36571841b3c0","Type":"ContainerStarted","Data":"99b849db1e9d702ce17759fc56ec702756241a9a61ca38bdef68a1003f351016"} Feb 25 13:48:48 crc kubenswrapper[4815]: I0225 13:48:48.935392 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:48:48 crc kubenswrapper[4815]: E0225 13:48:48.935853 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:48:49 crc kubenswrapper[4815]: I0225 13:48:49.724193 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" event={"ID":"d6add50b-b148-4e9b-af2a-36571841b3c0","Type":"ContainerStarted","Data":"b562a12088edf316046e99f2b810668a9b3289feff69bb0254372febfb724ead"} Feb 25 13:48:49 crc kubenswrapper[4815]: I0225 13:48:49.748475 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" podStartSLOduration=2.3019507470000002 podStartE2EDuration="2.748457457s" podCreationTimestamp="2026-02-25 13:48:47 +0000 UTC" firstStartedPulling="2026-02-25 13:48:48.662064244 +0000 UTC m=+1686.463162298" lastFinishedPulling="2026-02-25 13:48:49.108570944 +0000 UTC m=+1686.909669008" observedRunningTime="2026-02-25 13:48:49.745543717 +0000 UTC m=+1687.546641791" watchObservedRunningTime="2026-02-25 13:48:49.748457457 +0000 UTC m=+1687.549555531" Feb 25 13:49:01 crc kubenswrapper[4815]: I0225 13:49:01.936441 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:49:01 crc kubenswrapper[4815]: E0225 13:49:01.938474 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:49:11 crc kubenswrapper[4815]: I0225 13:49:11.057950 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-gcl8c"] Feb 25 13:49:11 crc kubenswrapper[4815]: I0225 13:49:11.072376 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-gcl8c"] Feb 25 13:49:12 crc kubenswrapper[4815]: I0225 13:49:12.948025 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:49:12 crc kubenswrapper[4815]: E0225 13:49:12.948873 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:49:12 crc kubenswrapper[4815]: I0225 13:49:12.950104 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2da1710b-375f-4795-993d-730ccbd04ccc" path="/var/lib/kubelet/pods/2da1710b-375f-4795-993d-730ccbd04ccc/volumes" Feb 25 13:49:13 crc kubenswrapper[4815]: I0225 13:49:13.044925 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-9337-account-create-update-86n94"] Feb 25 13:49:13 crc kubenswrapper[4815]: I0225 13:49:13.061152 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-5784-account-create-update-4gwt4"] Feb 25 13:49:13 crc kubenswrapper[4815]: I0225 13:49:13.077921 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0d2e-account-create-update-c7zn7"] Feb 25 13:49:13 crc kubenswrapper[4815]: I0225 13:49:13.087866 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-jldcn"] Feb 25 13:49:13 crc kubenswrapper[4815]: I0225 13:49:13.097803 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-9337-account-create-update-86n94"] Feb 25 13:49:13 crc kubenswrapper[4815]: I0225 13:49:13.107536 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-p95dh"] Feb 25 13:49:13 crc kubenswrapper[4815]: I0225 13:49:13.118528 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-jldcn"] Feb 25 13:49:13 crc kubenswrapper[4815]: I0225 13:49:13.126873 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0d2e-account-create-update-c7zn7"] Feb 25 13:49:13 crc kubenswrapper[4815]: I0225 13:49:13.139139 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-5784-account-create-update-4gwt4"] Feb 25 13:49:13 crc kubenswrapper[4815]: I0225 13:49:13.147214 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-p95dh"] Feb 25 13:49:14 crc kubenswrapper[4815]: I0225 13:49:14.946218 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="036fbace-3dcf-484d-958d-10970b9c7177" path="/var/lib/kubelet/pods/036fbace-3dcf-484d-958d-10970b9c7177/volumes" Feb 25 13:49:14 crc kubenswrapper[4815]: I0225 13:49:14.946764 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="486b281d-a6fc-4c69-860a-ad0d5480fc77" path="/var/lib/kubelet/pods/486b281d-a6fc-4c69-860a-ad0d5480fc77/volumes" Feb 25 13:49:14 crc kubenswrapper[4815]: I0225 13:49:14.947250 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6648a53c-a316-42aa-8799-8a1decdb704c" path="/var/lib/kubelet/pods/6648a53c-a316-42aa-8799-8a1decdb704c/volumes" Feb 25 13:49:14 crc kubenswrapper[4815]: I0225 13:49:14.947749 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="797bc888-260e-418d-b8a7-e0ff2b48a9af" path="/var/lib/kubelet/pods/797bc888-260e-418d-b8a7-e0ff2b48a9af/volumes" Feb 25 13:49:14 crc kubenswrapper[4815]: I0225 13:49:14.948665 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d139922-3032-4e41-978e-6bdd8ad8e0e1" path="/var/lib/kubelet/pods/7d139922-3032-4e41-978e-6bdd8ad8e0e1/volumes" Feb 25 13:49:24 crc kubenswrapper[4815]: I0225 13:49:24.066082 4815 generic.go:334] "Generic (PLEG): container finished" podID="d6add50b-b148-4e9b-af2a-36571841b3c0" containerID="b562a12088edf316046e99f2b810668a9b3289feff69bb0254372febfb724ead" exitCode=0 Feb 25 13:49:24 crc kubenswrapper[4815]: I0225 13:49:24.066163 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" event={"ID":"d6add50b-b148-4e9b-af2a-36571841b3c0","Type":"ContainerDied","Data":"b562a12088edf316046e99f2b810668a9b3289feff69bb0254372febfb724ead"} Feb 25 13:49:25 crc kubenswrapper[4815]: I0225 13:49:25.534625 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" Feb 25 13:49:25 crc kubenswrapper[4815]: I0225 13:49:25.687194 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6add50b-b148-4e9b-af2a-36571841b3c0-inventory\") pod \"d6add50b-b148-4e9b-af2a-36571841b3c0\" (UID: \"d6add50b-b148-4e9b-af2a-36571841b3c0\") " Feb 25 13:49:25 crc kubenswrapper[4815]: I0225 13:49:25.687952 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d6add50b-b148-4e9b-af2a-36571841b3c0-ssh-key-openstack-edpm-ipam\") pod \"d6add50b-b148-4e9b-af2a-36571841b3c0\" (UID: \"d6add50b-b148-4e9b-af2a-36571841b3c0\") " Feb 25 13:49:25 crc kubenswrapper[4815]: I0225 13:49:25.688087 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9vxd\" (UniqueName: \"kubernetes.io/projected/d6add50b-b148-4e9b-af2a-36571841b3c0-kube-api-access-d9vxd\") pod \"d6add50b-b148-4e9b-af2a-36571841b3c0\" (UID: \"d6add50b-b148-4e9b-af2a-36571841b3c0\") " Feb 25 13:49:25 crc kubenswrapper[4815]: I0225 13:49:25.700815 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6add50b-b148-4e9b-af2a-36571841b3c0-kube-api-access-d9vxd" (OuterVolumeSpecName: "kube-api-access-d9vxd") pod "d6add50b-b148-4e9b-af2a-36571841b3c0" (UID: "d6add50b-b148-4e9b-af2a-36571841b3c0"). InnerVolumeSpecName "kube-api-access-d9vxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:49:25 crc kubenswrapper[4815]: I0225 13:49:25.717802 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6add50b-b148-4e9b-af2a-36571841b3c0-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d6add50b-b148-4e9b-af2a-36571841b3c0" (UID: "d6add50b-b148-4e9b-af2a-36571841b3c0"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:49:25 crc kubenswrapper[4815]: I0225 13:49:25.732411 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6add50b-b148-4e9b-af2a-36571841b3c0-inventory" (OuterVolumeSpecName: "inventory") pod "d6add50b-b148-4e9b-af2a-36571841b3c0" (UID: "d6add50b-b148-4e9b-af2a-36571841b3c0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:49:25 crc kubenswrapper[4815]: I0225 13:49:25.790793 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6add50b-b148-4e9b-af2a-36571841b3c0-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 13:49:25 crc kubenswrapper[4815]: I0225 13:49:25.790849 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d6add50b-b148-4e9b-af2a-36571841b3c0-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 13:49:25 crc kubenswrapper[4815]: I0225 13:49:25.790866 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9vxd\" (UniqueName: \"kubernetes.io/projected/d6add50b-b148-4e9b-af2a-36571841b3c0-kube-api-access-d9vxd\") on node \"crc\" DevicePath \"\"" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.086335 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" event={"ID":"d6add50b-b148-4e9b-af2a-36571841b3c0","Type":"ContainerDied","Data":"99b849db1e9d702ce17759fc56ec702756241a9a61ca38bdef68a1003f351016"} Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.086399 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99b849db1e9d702ce17759fc56ec702756241a9a61ca38bdef68a1003f351016" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.086424 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sgwbs" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.180553 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb"] Feb 25 13:49:26 crc kubenswrapper[4815]: E0225 13:49:26.181294 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6add50b-b148-4e9b-af2a-36571841b3c0" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.181317 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6add50b-b148-4e9b-af2a-36571841b3c0" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.181574 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6add50b-b148-4e9b-af2a-36571841b3c0" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.182253 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.187176 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.187199 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.187321 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.187373 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mmbj" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.196737 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb"] Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.300879 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9f769896-45d8-45d0-8629-cee07d2cf4ea-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fpslb\" (UID: \"9f769896-45d8-45d0-8629-cee07d2cf4ea\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.301281 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f769896-45d8-45d0-8629-cee07d2cf4ea-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fpslb\" (UID: \"9f769896-45d8-45d0-8629-cee07d2cf4ea\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.301561 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p48b\" (UniqueName: \"kubernetes.io/projected/9f769896-45d8-45d0-8629-cee07d2cf4ea-kube-api-access-6p48b\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fpslb\" (UID: \"9f769896-45d8-45d0-8629-cee07d2cf4ea\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.404143 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9f769896-45d8-45d0-8629-cee07d2cf4ea-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fpslb\" (UID: \"9f769896-45d8-45d0-8629-cee07d2cf4ea\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.404636 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f769896-45d8-45d0-8629-cee07d2cf4ea-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fpslb\" (UID: \"9f769896-45d8-45d0-8629-cee07d2cf4ea\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.404778 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p48b\" (UniqueName: \"kubernetes.io/projected/9f769896-45d8-45d0-8629-cee07d2cf4ea-kube-api-access-6p48b\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fpslb\" (UID: \"9f769896-45d8-45d0-8629-cee07d2cf4ea\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.408631 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9f769896-45d8-45d0-8629-cee07d2cf4ea-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fpslb\" (UID: \"9f769896-45d8-45d0-8629-cee07d2cf4ea\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.411348 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f769896-45d8-45d0-8629-cee07d2cf4ea-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fpslb\" (UID: \"9f769896-45d8-45d0-8629-cee07d2cf4ea\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.423061 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p48b\" (UniqueName: \"kubernetes.io/projected/9f769896-45d8-45d0-8629-cee07d2cf4ea-kube-api-access-6p48b\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fpslb\" (UID: \"9f769896-45d8-45d0-8629-cee07d2cf4ea\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" Feb 25 13:49:26 crc kubenswrapper[4815]: I0225 13:49:26.515140 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" Feb 25 13:49:27 crc kubenswrapper[4815]: I0225 13:49:27.067141 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb"] Feb 25 13:49:27 crc kubenswrapper[4815]: I0225 13:49:27.070857 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 13:49:27 crc kubenswrapper[4815]: I0225 13:49:27.096814 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" event={"ID":"9f769896-45d8-45d0-8629-cee07d2cf4ea","Type":"ContainerStarted","Data":"428fe567485ff1a10b5960137fdf996e7699cb1f5e2eac56a264945f1c91865f"} Feb 25 13:49:27 crc kubenswrapper[4815]: I0225 13:49:27.935393 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:49:27 crc kubenswrapper[4815]: E0225 13:49:27.935747 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:49:28 crc kubenswrapper[4815]: I0225 13:49:28.107861 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" event={"ID":"9f769896-45d8-45d0-8629-cee07d2cf4ea","Type":"ContainerStarted","Data":"0f64f70924c2b6719e797efbf6535b4dc09c572558a98ad0afc4f112212a8552"} Feb 25 13:49:28 crc kubenswrapper[4815]: I0225 13:49:28.135141 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" podStartSLOduration=1.35096372 podStartE2EDuration="2.135114826s" podCreationTimestamp="2026-02-25 13:49:26 +0000 UTC" firstStartedPulling="2026-02-25 13:49:27.07058859 +0000 UTC m=+1724.871686664" lastFinishedPulling="2026-02-25 13:49:27.854739686 +0000 UTC m=+1725.655837770" observedRunningTime="2026-02-25 13:49:28.125017312 +0000 UTC m=+1725.926115376" watchObservedRunningTime="2026-02-25 13:49:28.135114826 +0000 UTC m=+1725.936212870" Feb 25 13:49:39 crc kubenswrapper[4815]: I0225 13:49:39.046027 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6t9rm"] Feb 25 13:49:39 crc kubenswrapper[4815]: I0225 13:49:39.055233 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6t9rm"] Feb 25 13:49:40 crc kubenswrapper[4815]: I0225 13:49:40.945483 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2580f26-4938-40dc-8a0f-044232a70469" path="/var/lib/kubelet/pods/e2580f26-4938-40dc-8a0f-044232a70469/volumes" Feb 25 13:49:41 crc kubenswrapper[4815]: I0225 13:49:41.936381 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:49:41 crc kubenswrapper[4815]: E0225 13:49:41.937171 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:49:47 crc kubenswrapper[4815]: I0225 13:49:47.381564 4815 scope.go:117] "RemoveContainer" containerID="584628a831021fe1c356deec621859ebe6086c452d288829b49333e39571b7b9" Feb 25 13:49:47 crc kubenswrapper[4815]: I0225 13:49:47.430541 4815 scope.go:117] "RemoveContainer" containerID="1e214a4fc38a0f6ae8dfc700a53a44d035a833bae6de51f606dc41c4523c7914" Feb 25 13:49:47 crc kubenswrapper[4815]: I0225 13:49:47.500811 4815 scope.go:117] "RemoveContainer" containerID="ce07eb1dae824de5b0f51fe31b9a37f69fba77d6ee95300dc5ae30b254ebf417" Feb 25 13:49:47 crc kubenswrapper[4815]: I0225 13:49:47.593213 4815 scope.go:117] "RemoveContainer" containerID="f48af9cccc860922855910f8e43bee585e03439414b4ad16b244780831db02b0" Feb 25 13:49:47 crc kubenswrapper[4815]: I0225 13:49:47.625175 4815 scope.go:117] "RemoveContainer" containerID="4ba8b1744c2f0247354c6cb552c320b03a8105617c24401eac941a74cdf5eb29" Feb 25 13:49:47 crc kubenswrapper[4815]: I0225 13:49:47.659594 4815 scope.go:117] "RemoveContainer" containerID="0d7ba17420be972376accbbcf2ccefc5fffda46d83ab793c0d542847d6d77abf" Feb 25 13:49:47 crc kubenswrapper[4815]: I0225 13:49:47.702833 4815 scope.go:117] "RemoveContainer" containerID="87466353d055e76f7d2aac41f39bf94d60ddd3053fbcf2d908e16d25074f648c" Feb 25 13:49:56 crc kubenswrapper[4815]: I0225 13:49:56.935532 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:49:56 crc kubenswrapper[4815]: E0225 13:49:56.936672 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:50:00 crc kubenswrapper[4815]: I0225 13:50:00.149193 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533790-6cjzj"] Feb 25 13:50:00 crc kubenswrapper[4815]: I0225 13:50:00.151437 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533790-6cjzj" Feb 25 13:50:00 crc kubenswrapper[4815]: I0225 13:50:00.153928 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:50:00 crc kubenswrapper[4815]: I0225 13:50:00.154150 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:50:00 crc kubenswrapper[4815]: I0225 13:50:00.154590 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:50:00 crc kubenswrapper[4815]: I0225 13:50:00.160782 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533790-6cjzj"] Feb 25 13:50:00 crc kubenswrapper[4815]: I0225 13:50:00.185464 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b44h4\" (UniqueName: \"kubernetes.io/projected/893e8504-9fb6-4734-aa82-aac7dfecb7c7-kube-api-access-b44h4\") pod \"auto-csr-approver-29533790-6cjzj\" (UID: \"893e8504-9fb6-4734-aa82-aac7dfecb7c7\") " pod="openshift-infra/auto-csr-approver-29533790-6cjzj" Feb 25 13:50:00 crc kubenswrapper[4815]: I0225 13:50:00.288076 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b44h4\" (UniqueName: \"kubernetes.io/projected/893e8504-9fb6-4734-aa82-aac7dfecb7c7-kube-api-access-b44h4\") pod \"auto-csr-approver-29533790-6cjzj\" (UID: \"893e8504-9fb6-4734-aa82-aac7dfecb7c7\") " pod="openshift-infra/auto-csr-approver-29533790-6cjzj" Feb 25 13:50:00 crc kubenswrapper[4815]: I0225 13:50:00.314204 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b44h4\" (UniqueName: \"kubernetes.io/projected/893e8504-9fb6-4734-aa82-aac7dfecb7c7-kube-api-access-b44h4\") pod \"auto-csr-approver-29533790-6cjzj\" (UID: \"893e8504-9fb6-4734-aa82-aac7dfecb7c7\") " pod="openshift-infra/auto-csr-approver-29533790-6cjzj" Feb 25 13:50:00 crc kubenswrapper[4815]: I0225 13:50:00.475639 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533790-6cjzj" Feb 25 13:50:00 crc kubenswrapper[4815]: I0225 13:50:00.924506 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533790-6cjzj"] Feb 25 13:50:01 crc kubenswrapper[4815]: I0225 13:50:01.437164 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533790-6cjzj" event={"ID":"893e8504-9fb6-4734-aa82-aac7dfecb7c7","Type":"ContainerStarted","Data":"e2a0a376a3fc6263f17b28809f3db04e3e303fd96b0105816f3653bf936423f8"} Feb 25 13:50:02 crc kubenswrapper[4815]: I0225 13:50:02.451406 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533790-6cjzj" event={"ID":"893e8504-9fb6-4734-aa82-aac7dfecb7c7","Type":"ContainerStarted","Data":"88339ec1ab335acc17c088dc3f0fd6d9ce7ab5947f12755f719e8a51056f92c1"} Feb 25 13:50:02 crc kubenswrapper[4815]: I0225 13:50:02.471964 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533790-6cjzj" podStartSLOduration=1.349399278 podStartE2EDuration="2.471945943s" podCreationTimestamp="2026-02-25 13:50:00 +0000 UTC" firstStartedPulling="2026-02-25 13:50:00.933913492 +0000 UTC m=+1758.735011546" lastFinishedPulling="2026-02-25 13:50:02.056460157 +0000 UTC m=+1759.857558211" observedRunningTime="2026-02-25 13:50:02.470399455 +0000 UTC m=+1760.271497539" watchObservedRunningTime="2026-02-25 13:50:02.471945943 +0000 UTC m=+1760.273043997" Feb 25 13:50:03 crc kubenswrapper[4815]: I0225 13:50:03.463589 4815 generic.go:334] "Generic (PLEG): container finished" podID="893e8504-9fb6-4734-aa82-aac7dfecb7c7" containerID="88339ec1ab335acc17c088dc3f0fd6d9ce7ab5947f12755f719e8a51056f92c1" exitCode=0 Feb 25 13:50:03 crc kubenswrapper[4815]: I0225 13:50:03.463680 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533790-6cjzj" event={"ID":"893e8504-9fb6-4734-aa82-aac7dfecb7c7","Type":"ContainerDied","Data":"88339ec1ab335acc17c088dc3f0fd6d9ce7ab5947f12755f719e8a51056f92c1"} Feb 25 13:50:04 crc kubenswrapper[4815]: I0225 13:50:04.857842 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533790-6cjzj" Feb 25 13:50:04 crc kubenswrapper[4815]: I0225 13:50:04.877929 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b44h4\" (UniqueName: \"kubernetes.io/projected/893e8504-9fb6-4734-aa82-aac7dfecb7c7-kube-api-access-b44h4\") pod \"893e8504-9fb6-4734-aa82-aac7dfecb7c7\" (UID: \"893e8504-9fb6-4734-aa82-aac7dfecb7c7\") " Feb 25 13:50:04 crc kubenswrapper[4815]: I0225 13:50:04.925685 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/893e8504-9fb6-4734-aa82-aac7dfecb7c7-kube-api-access-b44h4" (OuterVolumeSpecName: "kube-api-access-b44h4") pod "893e8504-9fb6-4734-aa82-aac7dfecb7c7" (UID: "893e8504-9fb6-4734-aa82-aac7dfecb7c7"). InnerVolumeSpecName "kube-api-access-b44h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:50:04 crc kubenswrapper[4815]: I0225 13:50:04.980690 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b44h4\" (UniqueName: \"kubernetes.io/projected/893e8504-9fb6-4734-aa82-aac7dfecb7c7-kube-api-access-b44h4\") on node \"crc\" DevicePath \"\"" Feb 25 13:50:05 crc kubenswrapper[4815]: I0225 13:50:05.483003 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533790-6cjzj" event={"ID":"893e8504-9fb6-4734-aa82-aac7dfecb7c7","Type":"ContainerDied","Data":"e2a0a376a3fc6263f17b28809f3db04e3e303fd96b0105816f3653bf936423f8"} Feb 25 13:50:05 crc kubenswrapper[4815]: I0225 13:50:05.483061 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2a0a376a3fc6263f17b28809f3db04e3e303fd96b0105816f3653bf936423f8" Feb 25 13:50:05 crc kubenswrapper[4815]: I0225 13:50:05.483089 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533790-6cjzj" Feb 25 13:50:05 crc kubenswrapper[4815]: I0225 13:50:05.542244 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533784-xcml6"] Feb 25 13:50:05 crc kubenswrapper[4815]: I0225 13:50:05.553940 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533784-xcml6"] Feb 25 13:50:06 crc kubenswrapper[4815]: I0225 13:50:06.947403 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad0b5715-982d-4406-9815-d281ee2d5ebf" path="/var/lib/kubelet/pods/ad0b5715-982d-4406-9815-d281ee2d5ebf/volumes" Feb 25 13:50:09 crc kubenswrapper[4815]: I0225 13:50:09.937380 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:50:09 crc kubenswrapper[4815]: E0225 13:50:09.938078 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:50:14 crc kubenswrapper[4815]: I0225 13:50:14.573353 4815 generic.go:334] "Generic (PLEG): container finished" podID="9f769896-45d8-45d0-8629-cee07d2cf4ea" containerID="0f64f70924c2b6719e797efbf6535b4dc09c572558a98ad0afc4f112212a8552" exitCode=0 Feb 25 13:50:14 crc kubenswrapper[4815]: I0225 13:50:14.573471 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" event={"ID":"9f769896-45d8-45d0-8629-cee07d2cf4ea","Type":"ContainerDied","Data":"0f64f70924c2b6719e797efbf6535b4dc09c572558a98ad0afc4f112212a8552"} Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.031299 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.227010 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9f769896-45d8-45d0-8629-cee07d2cf4ea-ssh-key-openstack-edpm-ipam\") pod \"9f769896-45d8-45d0-8629-cee07d2cf4ea\" (UID: \"9f769896-45d8-45d0-8629-cee07d2cf4ea\") " Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.227051 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f769896-45d8-45d0-8629-cee07d2cf4ea-inventory\") pod \"9f769896-45d8-45d0-8629-cee07d2cf4ea\" (UID: \"9f769896-45d8-45d0-8629-cee07d2cf4ea\") " Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.227192 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p48b\" (UniqueName: \"kubernetes.io/projected/9f769896-45d8-45d0-8629-cee07d2cf4ea-kube-api-access-6p48b\") pod \"9f769896-45d8-45d0-8629-cee07d2cf4ea\" (UID: \"9f769896-45d8-45d0-8629-cee07d2cf4ea\") " Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.233981 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f769896-45d8-45d0-8629-cee07d2cf4ea-kube-api-access-6p48b" (OuterVolumeSpecName: "kube-api-access-6p48b") pod "9f769896-45d8-45d0-8629-cee07d2cf4ea" (UID: "9f769896-45d8-45d0-8629-cee07d2cf4ea"). InnerVolumeSpecName "kube-api-access-6p48b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.255797 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f769896-45d8-45d0-8629-cee07d2cf4ea-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9f769896-45d8-45d0-8629-cee07d2cf4ea" (UID: "9f769896-45d8-45d0-8629-cee07d2cf4ea"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.278105 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f769896-45d8-45d0-8629-cee07d2cf4ea-inventory" (OuterVolumeSpecName: "inventory") pod "9f769896-45d8-45d0-8629-cee07d2cf4ea" (UID: "9f769896-45d8-45d0-8629-cee07d2cf4ea"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.331979 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9f769896-45d8-45d0-8629-cee07d2cf4ea-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.332587 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f769896-45d8-45d0-8629-cee07d2cf4ea-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.332624 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p48b\" (UniqueName: \"kubernetes.io/projected/9f769896-45d8-45d0-8629-cee07d2cf4ea-kube-api-access-6p48b\") on node \"crc\" DevicePath \"\"" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.599284 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" event={"ID":"9f769896-45d8-45d0-8629-cee07d2cf4ea","Type":"ContainerDied","Data":"428fe567485ff1a10b5960137fdf996e7699cb1f5e2eac56a264945f1c91865f"} Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.599358 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fpslb" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.599375 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="428fe567485ff1a10b5960137fdf996e7699cb1f5e2eac56a264945f1c91865f" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.701419 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-lfl6v"] Feb 25 13:50:16 crc kubenswrapper[4815]: E0225 13:50:16.701908 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="893e8504-9fb6-4734-aa82-aac7dfecb7c7" containerName="oc" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.701931 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="893e8504-9fb6-4734-aa82-aac7dfecb7c7" containerName="oc" Feb 25 13:50:16 crc kubenswrapper[4815]: E0225 13:50:16.701975 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f769896-45d8-45d0-8629-cee07d2cf4ea" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.701984 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f769896-45d8-45d0-8629-cee07d2cf4ea" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.702206 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f769896-45d8-45d0-8629-cee07d2cf4ea" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.702244 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="893e8504-9fb6-4734-aa82-aac7dfecb7c7" containerName="oc" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.703034 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.705606 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mmbj" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.705645 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.705915 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.707115 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.711824 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-lfl6v"] Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.740117 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/80323484-d533-4ca8-bfa5-315d318053f5-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-lfl6v\" (UID: \"80323484-d533-4ca8-bfa5-315d318053f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.740221 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/80323484-d533-4ca8-bfa5-315d318053f5-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-lfl6v\" (UID: \"80323484-d533-4ca8-bfa5-315d318053f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.740317 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qflcr\" (UniqueName: \"kubernetes.io/projected/80323484-d533-4ca8-bfa5-315d318053f5-kube-api-access-qflcr\") pod \"ssh-known-hosts-edpm-deployment-lfl6v\" (UID: \"80323484-d533-4ca8-bfa5-315d318053f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.842849 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/80323484-d533-4ca8-bfa5-315d318053f5-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-lfl6v\" (UID: \"80323484-d533-4ca8-bfa5-315d318053f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.842931 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/80323484-d533-4ca8-bfa5-315d318053f5-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-lfl6v\" (UID: \"80323484-d533-4ca8-bfa5-315d318053f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.842995 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qflcr\" (UniqueName: \"kubernetes.io/projected/80323484-d533-4ca8-bfa5-315d318053f5-kube-api-access-qflcr\") pod \"ssh-known-hosts-edpm-deployment-lfl6v\" (UID: \"80323484-d533-4ca8-bfa5-315d318053f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.847898 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/80323484-d533-4ca8-bfa5-315d318053f5-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-lfl6v\" (UID: \"80323484-d533-4ca8-bfa5-315d318053f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.848158 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/80323484-d533-4ca8-bfa5-315d318053f5-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-lfl6v\" (UID: \"80323484-d533-4ca8-bfa5-315d318053f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" Feb 25 13:50:16 crc kubenswrapper[4815]: I0225 13:50:16.873235 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qflcr\" (UniqueName: \"kubernetes.io/projected/80323484-d533-4ca8-bfa5-315d318053f5-kube-api-access-qflcr\") pod \"ssh-known-hosts-edpm-deployment-lfl6v\" (UID: \"80323484-d533-4ca8-bfa5-315d318053f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" Feb 25 13:50:17 crc kubenswrapper[4815]: I0225 13:50:17.019275 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" Feb 25 13:50:17 crc kubenswrapper[4815]: I0225 13:50:17.655658 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-lfl6v"] Feb 25 13:50:18 crc kubenswrapper[4815]: I0225 13:50:18.624185 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" event={"ID":"80323484-d533-4ca8-bfa5-315d318053f5","Type":"ContainerStarted","Data":"cbf670c09a1ee606883de96fb5a25b728b0dae001d3d06ab690911669f537ab5"} Feb 25 13:50:18 crc kubenswrapper[4815]: I0225 13:50:18.625234 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" event={"ID":"80323484-d533-4ca8-bfa5-315d318053f5","Type":"ContainerStarted","Data":"188f215475e9d3d4cba3f48f20c77e8bf0bbef245e499b77a7d3120ec8b5af83"} Feb 25 13:50:18 crc kubenswrapper[4815]: I0225 13:50:18.648294 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" podStartSLOduration=2.072149208 podStartE2EDuration="2.648233395s" podCreationTimestamp="2026-02-25 13:50:16 +0000 UTC" firstStartedPulling="2026-02-25 13:50:17.667926941 +0000 UTC m=+1775.469024995" lastFinishedPulling="2026-02-25 13:50:18.244011118 +0000 UTC m=+1776.045109182" observedRunningTime="2026-02-25 13:50:18.64423924 +0000 UTC m=+1776.445337314" watchObservedRunningTime="2026-02-25 13:50:18.648233395 +0000 UTC m=+1776.449331469" Feb 25 13:50:21 crc kubenswrapper[4815]: I0225 13:50:21.935535 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:50:21 crc kubenswrapper[4815]: E0225 13:50:21.936188 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:50:26 crc kubenswrapper[4815]: I0225 13:50:26.721805 4815 generic.go:334] "Generic (PLEG): container finished" podID="80323484-d533-4ca8-bfa5-315d318053f5" containerID="cbf670c09a1ee606883de96fb5a25b728b0dae001d3d06ab690911669f537ab5" exitCode=0 Feb 25 13:50:26 crc kubenswrapper[4815]: I0225 13:50:26.721908 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" event={"ID":"80323484-d533-4ca8-bfa5-315d318053f5","Type":"ContainerDied","Data":"cbf670c09a1ee606883de96fb5a25b728b0dae001d3d06ab690911669f537ab5"} Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.193080 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.343180 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/80323484-d533-4ca8-bfa5-315d318053f5-inventory-0\") pod \"80323484-d533-4ca8-bfa5-315d318053f5\" (UID: \"80323484-d533-4ca8-bfa5-315d318053f5\") " Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.343237 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qflcr\" (UniqueName: \"kubernetes.io/projected/80323484-d533-4ca8-bfa5-315d318053f5-kube-api-access-qflcr\") pod \"80323484-d533-4ca8-bfa5-315d318053f5\" (UID: \"80323484-d533-4ca8-bfa5-315d318053f5\") " Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.343313 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/80323484-d533-4ca8-bfa5-315d318053f5-ssh-key-openstack-edpm-ipam\") pod \"80323484-d533-4ca8-bfa5-315d318053f5\" (UID: \"80323484-d533-4ca8-bfa5-315d318053f5\") " Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.357439 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80323484-d533-4ca8-bfa5-315d318053f5-kube-api-access-qflcr" (OuterVolumeSpecName: "kube-api-access-qflcr") pod "80323484-d533-4ca8-bfa5-315d318053f5" (UID: "80323484-d533-4ca8-bfa5-315d318053f5"). InnerVolumeSpecName "kube-api-access-qflcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.387403 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80323484-d533-4ca8-bfa5-315d318053f5-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "80323484-d533-4ca8-bfa5-315d318053f5" (UID: "80323484-d533-4ca8-bfa5-315d318053f5"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.394093 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80323484-d533-4ca8-bfa5-315d318053f5-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "80323484-d533-4ca8-bfa5-315d318053f5" (UID: "80323484-d533-4ca8-bfa5-315d318053f5"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.446109 4815 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/80323484-d533-4ca8-bfa5-315d318053f5-inventory-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.446152 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qflcr\" (UniqueName: \"kubernetes.io/projected/80323484-d533-4ca8-bfa5-315d318053f5-kube-api-access-qflcr\") on node \"crc\" DevicePath \"\"" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.446167 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/80323484-d533-4ca8-bfa5-315d318053f5-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.749124 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" event={"ID":"80323484-d533-4ca8-bfa5-315d318053f5","Type":"ContainerDied","Data":"188f215475e9d3d4cba3f48f20c77e8bf0bbef245e499b77a7d3120ec8b5af83"} Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.749194 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="188f215475e9d3d4cba3f48f20c77e8bf0bbef245e499b77a7d3120ec8b5af83" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.749307 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-lfl6v" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.826841 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf"] Feb 25 13:50:28 crc kubenswrapper[4815]: E0225 13:50:28.827323 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80323484-d533-4ca8-bfa5-315d318053f5" containerName="ssh-known-hosts-edpm-deployment" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.827355 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="80323484-d533-4ca8-bfa5-315d318053f5" containerName="ssh-known-hosts-edpm-deployment" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.827659 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="80323484-d533-4ca8-bfa5-315d318053f5" containerName="ssh-known-hosts-edpm-deployment" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.828551 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.832642 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.832801 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mmbj" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.833197 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.834198 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.843982 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf"] Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.957349 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56gd8\" (UniqueName: \"kubernetes.io/projected/c037c85e-2d64-428a-a7c7-bcdabbb57676-kube-api-access-56gd8\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkxrf\" (UID: \"c037c85e-2d64-428a-a7c7-bcdabbb57676\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.957581 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c037c85e-2d64-428a-a7c7-bcdabbb57676-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkxrf\" (UID: \"c037c85e-2d64-428a-a7c7-bcdabbb57676\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" Feb 25 13:50:28 crc kubenswrapper[4815]: I0225 13:50:28.957705 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c037c85e-2d64-428a-a7c7-bcdabbb57676-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkxrf\" (UID: \"c037c85e-2d64-428a-a7c7-bcdabbb57676\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" Feb 25 13:50:29 crc kubenswrapper[4815]: I0225 13:50:29.059918 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c037c85e-2d64-428a-a7c7-bcdabbb57676-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkxrf\" (UID: \"c037c85e-2d64-428a-a7c7-bcdabbb57676\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" Feb 25 13:50:29 crc kubenswrapper[4815]: I0225 13:50:29.060049 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56gd8\" (UniqueName: \"kubernetes.io/projected/c037c85e-2d64-428a-a7c7-bcdabbb57676-kube-api-access-56gd8\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkxrf\" (UID: \"c037c85e-2d64-428a-a7c7-bcdabbb57676\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" Feb 25 13:50:29 crc kubenswrapper[4815]: I0225 13:50:29.062064 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c037c85e-2d64-428a-a7c7-bcdabbb57676-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkxrf\" (UID: \"c037c85e-2d64-428a-a7c7-bcdabbb57676\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" Feb 25 13:50:29 crc kubenswrapper[4815]: I0225 13:50:29.068297 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c037c85e-2d64-428a-a7c7-bcdabbb57676-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkxrf\" (UID: \"c037c85e-2d64-428a-a7c7-bcdabbb57676\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" Feb 25 13:50:29 crc kubenswrapper[4815]: I0225 13:50:29.068323 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c037c85e-2d64-428a-a7c7-bcdabbb57676-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkxrf\" (UID: \"c037c85e-2d64-428a-a7c7-bcdabbb57676\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" Feb 25 13:50:29 crc kubenswrapper[4815]: I0225 13:50:29.090934 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56gd8\" (UniqueName: \"kubernetes.io/projected/c037c85e-2d64-428a-a7c7-bcdabbb57676-kube-api-access-56gd8\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xkxrf\" (UID: \"c037c85e-2d64-428a-a7c7-bcdabbb57676\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" Feb 25 13:50:29 crc kubenswrapper[4815]: I0225 13:50:29.156135 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" Feb 25 13:50:29 crc kubenswrapper[4815]: I0225 13:50:29.709226 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf"] Feb 25 13:50:29 crc kubenswrapper[4815]: I0225 13:50:29.762535 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" event={"ID":"c037c85e-2d64-428a-a7c7-bcdabbb57676","Type":"ContainerStarted","Data":"13132da2e7c15b7bdd2a2b91422f67ab33f46113a9b58abe67c3ce0f9fae7635"} Feb 25 13:50:30 crc kubenswrapper[4815]: I0225 13:50:30.775331 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" event={"ID":"c037c85e-2d64-428a-a7c7-bcdabbb57676","Type":"ContainerStarted","Data":"d999ba533ea4a967d5adf5239be8b5538de080cfa314ffb0324180fbded07262"} Feb 25 13:50:30 crc kubenswrapper[4815]: I0225 13:50:30.796098 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" podStartSLOduration=2.368435983 podStartE2EDuration="2.796081117s" podCreationTimestamp="2026-02-25 13:50:28 +0000 UTC" firstStartedPulling="2026-02-25 13:50:29.71570522 +0000 UTC m=+1787.516803284" lastFinishedPulling="2026-02-25 13:50:30.143350354 +0000 UTC m=+1787.944448418" observedRunningTime="2026-02-25 13:50:30.791159305 +0000 UTC m=+1788.592257359" watchObservedRunningTime="2026-02-25 13:50:30.796081117 +0000 UTC m=+1788.597179171" Feb 25 13:50:34 crc kubenswrapper[4815]: I0225 13:50:34.033764 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-xkz82"] Feb 25 13:50:34 crc kubenswrapper[4815]: I0225 13:50:34.049578 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-xkz82"] Feb 25 13:50:34 crc kubenswrapper[4815]: I0225 13:50:34.061002 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5n7jw"] Feb 25 13:50:34 crc kubenswrapper[4815]: I0225 13:50:34.067838 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5n7jw"] Feb 25 13:50:34 crc kubenswrapper[4815]: I0225 13:50:34.947130 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="947c26ca-d609-4a2a-a4d8-0ab4ae3f69be" path="/var/lib/kubelet/pods/947c26ca-d609-4a2a-a4d8-0ab4ae3f69be/volumes" Feb 25 13:50:34 crc kubenswrapper[4815]: I0225 13:50:34.947939 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98fc77c3-6063-41a5-ba29-0875703dfe05" path="/var/lib/kubelet/pods/98fc77c3-6063-41a5-ba29-0875703dfe05/volumes" Feb 25 13:50:36 crc kubenswrapper[4815]: I0225 13:50:36.935964 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:50:36 crc kubenswrapper[4815]: E0225 13:50:36.936634 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:50:38 crc kubenswrapper[4815]: I0225 13:50:38.855856 4815 generic.go:334] "Generic (PLEG): container finished" podID="c037c85e-2d64-428a-a7c7-bcdabbb57676" containerID="d999ba533ea4a967d5adf5239be8b5538de080cfa314ffb0324180fbded07262" exitCode=0 Feb 25 13:50:38 crc kubenswrapper[4815]: I0225 13:50:38.856752 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" event={"ID":"c037c85e-2d64-428a-a7c7-bcdabbb57676","Type":"ContainerDied","Data":"d999ba533ea4a967d5adf5239be8b5538de080cfa314ffb0324180fbded07262"} Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.304548 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.402474 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c037c85e-2d64-428a-a7c7-bcdabbb57676-ssh-key-openstack-edpm-ipam\") pod \"c037c85e-2d64-428a-a7c7-bcdabbb57676\" (UID: \"c037c85e-2d64-428a-a7c7-bcdabbb57676\") " Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.402583 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c037c85e-2d64-428a-a7c7-bcdabbb57676-inventory\") pod \"c037c85e-2d64-428a-a7c7-bcdabbb57676\" (UID: \"c037c85e-2d64-428a-a7c7-bcdabbb57676\") " Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.402991 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56gd8\" (UniqueName: \"kubernetes.io/projected/c037c85e-2d64-428a-a7c7-bcdabbb57676-kube-api-access-56gd8\") pod \"c037c85e-2d64-428a-a7c7-bcdabbb57676\" (UID: \"c037c85e-2d64-428a-a7c7-bcdabbb57676\") " Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.409727 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c037c85e-2d64-428a-a7c7-bcdabbb57676-kube-api-access-56gd8" (OuterVolumeSpecName: "kube-api-access-56gd8") pod "c037c85e-2d64-428a-a7c7-bcdabbb57676" (UID: "c037c85e-2d64-428a-a7c7-bcdabbb57676"). InnerVolumeSpecName "kube-api-access-56gd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.431460 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c037c85e-2d64-428a-a7c7-bcdabbb57676-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c037c85e-2d64-428a-a7c7-bcdabbb57676" (UID: "c037c85e-2d64-428a-a7c7-bcdabbb57676"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.453106 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c037c85e-2d64-428a-a7c7-bcdabbb57676-inventory" (OuterVolumeSpecName: "inventory") pod "c037c85e-2d64-428a-a7c7-bcdabbb57676" (UID: "c037c85e-2d64-428a-a7c7-bcdabbb57676"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.505758 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56gd8\" (UniqueName: \"kubernetes.io/projected/c037c85e-2d64-428a-a7c7-bcdabbb57676-kube-api-access-56gd8\") on node \"crc\" DevicePath \"\"" Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.505797 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c037c85e-2d64-428a-a7c7-bcdabbb57676-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.505808 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c037c85e-2d64-428a-a7c7-bcdabbb57676-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.877595 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" event={"ID":"c037c85e-2d64-428a-a7c7-bcdabbb57676","Type":"ContainerDied","Data":"13132da2e7c15b7bdd2a2b91422f67ab33f46113a9b58abe67c3ce0f9fae7635"} Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.877636 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13132da2e7c15b7bdd2a2b91422f67ab33f46113a9b58abe67c3ce0f9fae7635" Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.877687 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xkxrf" Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.972340 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866"] Feb 25 13:50:40 crc kubenswrapper[4815]: E0225 13:50:40.972910 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c037c85e-2d64-428a-a7c7-bcdabbb57676" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.972939 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="c037c85e-2d64-428a-a7c7-bcdabbb57676" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.973251 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="c037c85e-2d64-428a-a7c7-bcdabbb57676" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.974197 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.982953 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.983044 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.983057 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.983121 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mmbj" Feb 25 13:50:40 crc kubenswrapper[4815]: I0225 13:50:40.984857 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866"] Feb 25 13:50:41 crc kubenswrapper[4815]: I0225 13:50:41.132172 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55562f1f-6c46-4413-9f01-99867d3a1a7f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tn866\" (UID: \"55562f1f-6c46-4413-9f01-99867d3a1a7f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" Feb 25 13:50:41 crc kubenswrapper[4815]: I0225 13:50:41.132310 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55562f1f-6c46-4413-9f01-99867d3a1a7f-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tn866\" (UID: \"55562f1f-6c46-4413-9f01-99867d3a1a7f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" Feb 25 13:50:41 crc kubenswrapper[4815]: I0225 13:50:41.132392 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgrtr\" (UniqueName: \"kubernetes.io/projected/55562f1f-6c46-4413-9f01-99867d3a1a7f-kube-api-access-bgrtr\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tn866\" (UID: \"55562f1f-6c46-4413-9f01-99867d3a1a7f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" Feb 25 13:50:41 crc kubenswrapper[4815]: I0225 13:50:41.234179 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgrtr\" (UniqueName: \"kubernetes.io/projected/55562f1f-6c46-4413-9f01-99867d3a1a7f-kube-api-access-bgrtr\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tn866\" (UID: \"55562f1f-6c46-4413-9f01-99867d3a1a7f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" Feb 25 13:50:41 crc kubenswrapper[4815]: I0225 13:50:41.234444 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55562f1f-6c46-4413-9f01-99867d3a1a7f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tn866\" (UID: \"55562f1f-6c46-4413-9f01-99867d3a1a7f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" Feb 25 13:50:41 crc kubenswrapper[4815]: I0225 13:50:41.234549 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55562f1f-6c46-4413-9f01-99867d3a1a7f-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tn866\" (UID: \"55562f1f-6c46-4413-9f01-99867d3a1a7f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" Feb 25 13:50:41 crc kubenswrapper[4815]: I0225 13:50:41.239621 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55562f1f-6c46-4413-9f01-99867d3a1a7f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tn866\" (UID: \"55562f1f-6c46-4413-9f01-99867d3a1a7f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" Feb 25 13:50:41 crc kubenswrapper[4815]: I0225 13:50:41.239674 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55562f1f-6c46-4413-9f01-99867d3a1a7f-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tn866\" (UID: \"55562f1f-6c46-4413-9f01-99867d3a1a7f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" Feb 25 13:50:41 crc kubenswrapper[4815]: I0225 13:50:41.250301 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgrtr\" (UniqueName: \"kubernetes.io/projected/55562f1f-6c46-4413-9f01-99867d3a1a7f-kube-api-access-bgrtr\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-tn866\" (UID: \"55562f1f-6c46-4413-9f01-99867d3a1a7f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" Feb 25 13:50:41 crc kubenswrapper[4815]: I0225 13:50:41.302727 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" Feb 25 13:50:41 crc kubenswrapper[4815]: I0225 13:50:41.817961 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866"] Feb 25 13:50:41 crc kubenswrapper[4815]: W0225 13:50:41.825561 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55562f1f_6c46_4413_9f01_99867d3a1a7f.slice/crio-3fa181d7e9aca018b87a146bb1ed828192495bcf85d32e4d64f9e604595490c1 WatchSource:0}: Error finding container 3fa181d7e9aca018b87a146bb1ed828192495bcf85d32e4d64f9e604595490c1: Status 404 returned error can't find the container with id 3fa181d7e9aca018b87a146bb1ed828192495bcf85d32e4d64f9e604595490c1 Feb 25 13:50:41 crc kubenswrapper[4815]: I0225 13:50:41.887417 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" event={"ID":"55562f1f-6c46-4413-9f01-99867d3a1a7f","Type":"ContainerStarted","Data":"3fa181d7e9aca018b87a146bb1ed828192495bcf85d32e4d64f9e604595490c1"} Feb 25 13:50:42 crc kubenswrapper[4815]: I0225 13:50:42.920839 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" event={"ID":"55562f1f-6c46-4413-9f01-99867d3a1a7f","Type":"ContainerStarted","Data":"5d4fef2239bffb3c87a9a6c2e0c26e60bbcbe858ca9f6e8566a8bc5115428688"} Feb 25 13:50:42 crc kubenswrapper[4815]: I0225 13:50:42.957690 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" podStartSLOduration=2.5299514419999998 podStartE2EDuration="2.957668547s" podCreationTimestamp="2026-02-25 13:50:40 +0000 UTC" firstStartedPulling="2026-02-25 13:50:41.828678063 +0000 UTC m=+1799.629776117" lastFinishedPulling="2026-02-25 13:50:42.256395178 +0000 UTC m=+1800.057493222" observedRunningTime="2026-02-25 13:50:42.944740226 +0000 UTC m=+1800.745838290" watchObservedRunningTime="2026-02-25 13:50:42.957668547 +0000 UTC m=+1800.758766611" Feb 25 13:50:47 crc kubenswrapper[4815]: I0225 13:50:47.818729 4815 scope.go:117] "RemoveContainer" containerID="3b3fadc2c41679dd6cbd8de18d1e4ee943756cb1e45c6b8ea7320aee646486d0" Feb 25 13:50:47 crc kubenswrapper[4815]: I0225 13:50:47.889374 4815 scope.go:117] "RemoveContainer" containerID="c9a93486d2ab0ff1a07236d6049b2378f1b98706e40d3b5bdace6972d3e041dc" Feb 25 13:50:47 crc kubenswrapper[4815]: I0225 13:50:47.932133 4815 scope.go:117] "RemoveContainer" containerID="53c66837ec255d39fab425ff76173ac7ffa456950e2714481fd0aa74c95dbe5e" Feb 25 13:50:49 crc kubenswrapper[4815]: I0225 13:50:49.936386 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:50:49 crc kubenswrapper[4815]: E0225 13:50:49.937099 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:50:52 crc kubenswrapper[4815]: I0225 13:50:52.012276 4815 generic.go:334] "Generic (PLEG): container finished" podID="55562f1f-6c46-4413-9f01-99867d3a1a7f" containerID="5d4fef2239bffb3c87a9a6c2e0c26e60bbcbe858ca9f6e8566a8bc5115428688" exitCode=0 Feb 25 13:50:52 crc kubenswrapper[4815]: I0225 13:50:52.012380 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" event={"ID":"55562f1f-6c46-4413-9f01-99867d3a1a7f","Type":"ContainerDied","Data":"5d4fef2239bffb3c87a9a6c2e0c26e60bbcbe858ca9f6e8566a8bc5115428688"} Feb 25 13:50:53 crc kubenswrapper[4815]: I0225 13:50:53.541521 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" Feb 25 13:50:53 crc kubenswrapper[4815]: I0225 13:50:53.605715 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55562f1f-6c46-4413-9f01-99867d3a1a7f-ssh-key-openstack-edpm-ipam\") pod \"55562f1f-6c46-4413-9f01-99867d3a1a7f\" (UID: \"55562f1f-6c46-4413-9f01-99867d3a1a7f\") " Feb 25 13:50:53 crc kubenswrapper[4815]: I0225 13:50:53.606078 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55562f1f-6c46-4413-9f01-99867d3a1a7f-inventory\") pod \"55562f1f-6c46-4413-9f01-99867d3a1a7f\" (UID: \"55562f1f-6c46-4413-9f01-99867d3a1a7f\") " Feb 25 13:50:53 crc kubenswrapper[4815]: I0225 13:50:53.606141 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgrtr\" (UniqueName: \"kubernetes.io/projected/55562f1f-6c46-4413-9f01-99867d3a1a7f-kube-api-access-bgrtr\") pod \"55562f1f-6c46-4413-9f01-99867d3a1a7f\" (UID: \"55562f1f-6c46-4413-9f01-99867d3a1a7f\") " Feb 25 13:50:53 crc kubenswrapper[4815]: I0225 13:50:53.612390 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55562f1f-6c46-4413-9f01-99867d3a1a7f-kube-api-access-bgrtr" (OuterVolumeSpecName: "kube-api-access-bgrtr") pod "55562f1f-6c46-4413-9f01-99867d3a1a7f" (UID: "55562f1f-6c46-4413-9f01-99867d3a1a7f"). InnerVolumeSpecName "kube-api-access-bgrtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:50:53 crc kubenswrapper[4815]: I0225 13:50:53.651898 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55562f1f-6c46-4413-9f01-99867d3a1a7f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "55562f1f-6c46-4413-9f01-99867d3a1a7f" (UID: "55562f1f-6c46-4413-9f01-99867d3a1a7f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:50:53 crc kubenswrapper[4815]: I0225 13:50:53.680063 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55562f1f-6c46-4413-9f01-99867d3a1a7f-inventory" (OuterVolumeSpecName: "inventory") pod "55562f1f-6c46-4413-9f01-99867d3a1a7f" (UID: "55562f1f-6c46-4413-9f01-99867d3a1a7f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:50:53 crc kubenswrapper[4815]: I0225 13:50:53.708112 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgrtr\" (UniqueName: \"kubernetes.io/projected/55562f1f-6c46-4413-9f01-99867d3a1a7f-kube-api-access-bgrtr\") on node \"crc\" DevicePath \"\"" Feb 25 13:50:53 crc kubenswrapper[4815]: I0225 13:50:53.708148 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55562f1f-6c46-4413-9f01-99867d3a1a7f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 13:50:53 crc kubenswrapper[4815]: I0225 13:50:53.708160 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55562f1f-6c46-4413-9f01-99867d3a1a7f-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.034199 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" event={"ID":"55562f1f-6c46-4413-9f01-99867d3a1a7f","Type":"ContainerDied","Data":"3fa181d7e9aca018b87a146bb1ed828192495bcf85d32e4d64f9e604595490c1"} Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.034259 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fa181d7e9aca018b87a146bb1ed828192495bcf85d32e4d64f9e604595490c1" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.034259 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-tn866" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.153544 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v"] Feb 25 13:50:54 crc kubenswrapper[4815]: E0225 13:50:54.153996 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55562f1f-6c46-4413-9f01-99867d3a1a7f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.154019 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="55562f1f-6c46-4413-9f01-99867d3a1a7f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.154292 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="55562f1f-6c46-4413-9f01-99867d3a1a7f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.155073 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.160483 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.161017 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.161352 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.161653 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.162147 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mmbj" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.162687 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.162935 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.163146 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.163992 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v"] Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.217159 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.217245 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.217308 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.217365 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.217398 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g85mw\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-kube-api-access-g85mw\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.217452 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.217535 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.217580 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.217612 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.217671 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.217751 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.217869 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.217952 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.218007 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.320101 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.320151 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.320183 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.320209 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.320225 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g85mw\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-kube-api-access-g85mw\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.320249 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.320277 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.320297 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.320311 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.320339 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.320373 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.320421 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.320450 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.320474 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.324570 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.325277 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.325723 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.326063 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.326404 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.327159 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.328165 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.328206 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.328995 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.332500 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.333105 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.333162 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.334811 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.338810 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g85mw\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-kube-api-access-g85mw\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:54 crc kubenswrapper[4815]: I0225 13:50:54.525391 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:50:55 crc kubenswrapper[4815]: W0225 13:50:55.092835 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb70de6b1_ba0c_482d_a1d6_3c0604a374c1.slice/crio-90d690ccc8b821279980754a32935ffa792cc9daad055dfd10ce05ca44ed0cda WatchSource:0}: Error finding container 90d690ccc8b821279980754a32935ffa792cc9daad055dfd10ce05ca44ed0cda: Status 404 returned error can't find the container with id 90d690ccc8b821279980754a32935ffa792cc9daad055dfd10ce05ca44ed0cda Feb 25 13:50:55 crc kubenswrapper[4815]: I0225 13:50:55.099387 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v"] Feb 25 13:50:56 crc kubenswrapper[4815]: I0225 13:50:56.060944 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" event={"ID":"b70de6b1-ba0c-482d-a1d6-3c0604a374c1","Type":"ContainerStarted","Data":"dfaac2d1bccef6b020313a3d6e446e43a730e519929e1522c9456926fdc7fba4"} Feb 25 13:50:56 crc kubenswrapper[4815]: I0225 13:50:56.061568 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" event={"ID":"b70de6b1-ba0c-482d-a1d6-3c0604a374c1","Type":"ContainerStarted","Data":"90d690ccc8b821279980754a32935ffa792cc9daad055dfd10ce05ca44ed0cda"} Feb 25 13:50:56 crc kubenswrapper[4815]: I0225 13:50:56.095011 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" podStartSLOduration=1.6417280220000001 podStartE2EDuration="2.094965408s" podCreationTimestamp="2026-02-25 13:50:54 +0000 UTC" firstStartedPulling="2026-02-25 13:50:55.096074369 +0000 UTC m=+1812.897172443" lastFinishedPulling="2026-02-25 13:50:55.549311735 +0000 UTC m=+1813.350409829" observedRunningTime="2026-02-25 13:50:56.082897284 +0000 UTC m=+1813.883995378" watchObservedRunningTime="2026-02-25 13:50:56.094965408 +0000 UTC m=+1813.896063472" Feb 25 13:51:01 crc kubenswrapper[4815]: I0225 13:51:01.937029 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:51:01 crc kubenswrapper[4815]: E0225 13:51:01.938440 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:51:14 crc kubenswrapper[4815]: I0225 13:51:14.936487 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:51:14 crc kubenswrapper[4815]: E0225 13:51:14.937506 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:51:20 crc kubenswrapper[4815]: I0225 13:51:20.065494 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-8bpf7"] Feb 25 13:51:20 crc kubenswrapper[4815]: I0225 13:51:20.106142 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-8bpf7"] Feb 25 13:51:20 crc kubenswrapper[4815]: I0225 13:51:20.945072 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a752cbdd-832b-4635-84af-03cecc4af21e" path="/var/lib/kubelet/pods/a752cbdd-832b-4635-84af-03cecc4af21e/volumes" Feb 25 13:51:27 crc kubenswrapper[4815]: I0225 13:51:27.936344 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:51:28 crc kubenswrapper[4815]: I0225 13:51:28.391074 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerStarted","Data":"bc1a87ba95ff2fe57f3edfcdf517b6e2b034a474d761b615e638c4a52c55574e"} Feb 25 13:51:32 crc kubenswrapper[4815]: I0225 13:51:32.439425 4815 generic.go:334] "Generic (PLEG): container finished" podID="b70de6b1-ba0c-482d-a1d6-3c0604a374c1" containerID="dfaac2d1bccef6b020313a3d6e446e43a730e519929e1522c9456926fdc7fba4" exitCode=0 Feb 25 13:51:32 crc kubenswrapper[4815]: I0225 13:51:32.439562 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" event={"ID":"b70de6b1-ba0c-482d-a1d6-3c0604a374c1","Type":"ContainerDied","Data":"dfaac2d1bccef6b020313a3d6e446e43a730e519929e1522c9456926fdc7fba4"} Feb 25 13:51:33 crc kubenswrapper[4815]: I0225 13:51:33.935280 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.068852 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.069217 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-ovn-default-certs-0\") pod \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.069316 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-bootstrap-combined-ca-bundle\") pod \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.069344 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-inventory\") pod \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.069397 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-ovn-combined-ca-bundle\") pod \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.069424 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.069466 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.069490 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-neutron-metadata-combined-ca-bundle\") pod \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.069542 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g85mw\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-kube-api-access-g85mw\") pod \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.069585 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-repo-setup-combined-ca-bundle\") pod \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.069611 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-ssh-key-openstack-edpm-ipam\") pod \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.069629 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-nova-combined-ca-bundle\") pod \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.069679 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-libvirt-combined-ca-bundle\") pod \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.069703 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-telemetry-combined-ca-bundle\") pod \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\" (UID: \"b70de6b1-ba0c-482d-a1d6-3c0604a374c1\") " Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.079687 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "b70de6b1-ba0c-482d-a1d6-3c0604a374c1" (UID: "b70de6b1-ba0c-482d-a1d6-3c0604a374c1"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.079776 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "b70de6b1-ba0c-482d-a1d6-3c0604a374c1" (UID: "b70de6b1-ba0c-482d-a1d6-3c0604a374c1"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.080743 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "b70de6b1-ba0c-482d-a1d6-3c0604a374c1" (UID: "b70de6b1-ba0c-482d-a1d6-3c0604a374c1"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.080826 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b70de6b1-ba0c-482d-a1d6-3c0604a374c1" (UID: "b70de6b1-ba0c-482d-a1d6-3c0604a374c1"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.080919 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-kube-api-access-g85mw" (OuterVolumeSpecName: "kube-api-access-g85mw") pod "b70de6b1-ba0c-482d-a1d6-3c0604a374c1" (UID: "b70de6b1-ba0c-482d-a1d6-3c0604a374c1"). InnerVolumeSpecName "kube-api-access-g85mw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.080971 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "b70de6b1-ba0c-482d-a1d6-3c0604a374c1" (UID: "b70de6b1-ba0c-482d-a1d6-3c0604a374c1"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.080991 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "b70de6b1-ba0c-482d-a1d6-3c0604a374c1" (UID: "b70de6b1-ba0c-482d-a1d6-3c0604a374c1"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.081189 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "b70de6b1-ba0c-482d-a1d6-3c0604a374c1" (UID: "b70de6b1-ba0c-482d-a1d6-3c0604a374c1"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.083123 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "b70de6b1-ba0c-482d-a1d6-3c0604a374c1" (UID: "b70de6b1-ba0c-482d-a1d6-3c0604a374c1"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.089709 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b70de6b1-ba0c-482d-a1d6-3c0604a374c1" (UID: "b70de6b1-ba0c-482d-a1d6-3c0604a374c1"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.092979 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "b70de6b1-ba0c-482d-a1d6-3c0604a374c1" (UID: "b70de6b1-ba0c-482d-a1d6-3c0604a374c1"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.093286 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b70de6b1-ba0c-482d-a1d6-3c0604a374c1" (UID: "b70de6b1-ba0c-482d-a1d6-3c0604a374c1"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.106277 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-inventory" (OuterVolumeSpecName: "inventory") pod "b70de6b1-ba0c-482d-a1d6-3c0604a374c1" (UID: "b70de6b1-ba0c-482d-a1d6-3c0604a374c1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.109290 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b70de6b1-ba0c-482d-a1d6-3c0604a374c1" (UID: "b70de6b1-ba0c-482d-a1d6-3c0604a374c1"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.174946 4815 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.174979 4815 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.174994 4815 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.175007 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.175016 4815 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.175027 4815 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.175038 4815 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.175047 4815 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.175057 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g85mw\" (UniqueName: \"kubernetes.io/projected/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-kube-api-access-g85mw\") on node \"crc\" DevicePath \"\"" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.175066 4815 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.175074 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.175082 4815 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.175090 4815 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.175098 4815 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b70de6b1-ba0c-482d-a1d6-3c0604a374c1-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.462549 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" event={"ID":"b70de6b1-ba0c-482d-a1d6-3c0604a374c1","Type":"ContainerDied","Data":"90d690ccc8b821279980754a32935ffa792cc9daad055dfd10ce05ca44ed0cda"} Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.462787 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90d690ccc8b821279980754a32935ffa792cc9daad055dfd10ce05ca44ed0cda" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.462671 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.569758 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95"] Feb 25 13:51:34 crc kubenswrapper[4815]: E0225 13:51:34.570162 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b70de6b1-ba0c-482d-a1d6-3c0604a374c1" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.570177 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b70de6b1-ba0c-482d-a1d6-3c0604a374c1" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.570363 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b70de6b1-ba0c-482d-a1d6-3c0604a374c1" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.570995 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.574788 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mmbj" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.574814 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.574882 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.574967 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.575539 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.587758 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95"] Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.685572 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jvq95\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.685672 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jvq95\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.685798 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jvq95\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.685923 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8d7f9d84-5806-4beb-9277-c77da4440527-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jvq95\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.686058 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv6rk\" (UniqueName: \"kubernetes.io/projected/8d7f9d84-5806-4beb-9277-c77da4440527-kube-api-access-tv6rk\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jvq95\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.787585 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jvq95\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.787652 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jvq95\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.787684 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jvq95\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.787714 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8d7f9d84-5806-4beb-9277-c77da4440527-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jvq95\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.787763 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv6rk\" (UniqueName: \"kubernetes.io/projected/8d7f9d84-5806-4beb-9277-c77da4440527-kube-api-access-tv6rk\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jvq95\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.790099 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8d7f9d84-5806-4beb-9277-c77da4440527-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jvq95\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.792909 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jvq95\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.792934 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jvq95\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.793236 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jvq95\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.814376 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv6rk\" (UniqueName: \"kubernetes.io/projected/8d7f9d84-5806-4beb-9277-c77da4440527-kube-api-access-tv6rk\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jvq95\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:51:34 crc kubenswrapper[4815]: I0225 13:51:34.951824 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:51:35 crc kubenswrapper[4815]: I0225 13:51:35.516269 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95"] Feb 25 13:51:36 crc kubenswrapper[4815]: I0225 13:51:36.486735 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" event={"ID":"8d7f9d84-5806-4beb-9277-c77da4440527","Type":"ContainerStarted","Data":"f792b2707476dd0d63b9759613de232448f750d926051d29c127874a9dc2497c"} Feb 25 13:51:36 crc kubenswrapper[4815]: I0225 13:51:36.487930 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" event={"ID":"8d7f9d84-5806-4beb-9277-c77da4440527","Type":"ContainerStarted","Data":"1184df3214a105434046f8f7ae756f6702bbc6d4442f6d09a0fa1fb431b0c5ce"} Feb 25 13:51:36 crc kubenswrapper[4815]: I0225 13:51:36.508993 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" podStartSLOduration=2.064741101 podStartE2EDuration="2.508971269s" podCreationTimestamp="2026-02-25 13:51:34 +0000 UTC" firstStartedPulling="2026-02-25 13:51:35.525365733 +0000 UTC m=+1853.326463787" lastFinishedPulling="2026-02-25 13:51:35.969595861 +0000 UTC m=+1853.770693955" observedRunningTime="2026-02-25 13:51:36.506821443 +0000 UTC m=+1854.307919547" watchObservedRunningTime="2026-02-25 13:51:36.508971269 +0000 UTC m=+1854.310069333" Feb 25 13:51:48 crc kubenswrapper[4815]: I0225 13:51:48.048187 4815 scope.go:117] "RemoveContainer" containerID="e97e5bb19afbf571975e89fe8a65df1145d57e6c6b03e8207d710fb992a8aa33" Feb 25 13:52:00 crc kubenswrapper[4815]: I0225 13:52:00.147082 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533792-b2jvf"] Feb 25 13:52:00 crc kubenswrapper[4815]: I0225 13:52:00.149099 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533792-b2jvf" Feb 25 13:52:00 crc kubenswrapper[4815]: I0225 13:52:00.151723 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:52:00 crc kubenswrapper[4815]: I0225 13:52:00.151953 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:52:00 crc kubenswrapper[4815]: I0225 13:52:00.153793 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:52:00 crc kubenswrapper[4815]: I0225 13:52:00.173229 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533792-b2jvf"] Feb 25 13:52:00 crc kubenswrapper[4815]: I0225 13:52:00.317700 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsfhn\" (UniqueName: \"kubernetes.io/projected/4449c14e-2d79-4ce8-ba74-4368a26f5d44-kube-api-access-nsfhn\") pod \"auto-csr-approver-29533792-b2jvf\" (UID: \"4449c14e-2d79-4ce8-ba74-4368a26f5d44\") " pod="openshift-infra/auto-csr-approver-29533792-b2jvf" Feb 25 13:52:00 crc kubenswrapper[4815]: I0225 13:52:00.419480 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsfhn\" (UniqueName: \"kubernetes.io/projected/4449c14e-2d79-4ce8-ba74-4368a26f5d44-kube-api-access-nsfhn\") pod \"auto-csr-approver-29533792-b2jvf\" (UID: \"4449c14e-2d79-4ce8-ba74-4368a26f5d44\") " pod="openshift-infra/auto-csr-approver-29533792-b2jvf" Feb 25 13:52:00 crc kubenswrapper[4815]: I0225 13:52:00.458258 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsfhn\" (UniqueName: \"kubernetes.io/projected/4449c14e-2d79-4ce8-ba74-4368a26f5d44-kube-api-access-nsfhn\") pod \"auto-csr-approver-29533792-b2jvf\" (UID: \"4449c14e-2d79-4ce8-ba74-4368a26f5d44\") " pod="openshift-infra/auto-csr-approver-29533792-b2jvf" Feb 25 13:52:00 crc kubenswrapper[4815]: I0225 13:52:00.475456 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533792-b2jvf" Feb 25 13:52:00 crc kubenswrapper[4815]: I0225 13:52:00.952030 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533792-b2jvf"] Feb 25 13:52:01 crc kubenswrapper[4815]: I0225 13:52:01.726603 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533792-b2jvf" event={"ID":"4449c14e-2d79-4ce8-ba74-4368a26f5d44","Type":"ContainerStarted","Data":"9f5c2836005f7d5f6293e71f661fce0ec3b5d1600821ca50bec91fe889e35e6a"} Feb 25 13:52:02 crc kubenswrapper[4815]: I0225 13:52:02.747150 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533792-b2jvf" event={"ID":"4449c14e-2d79-4ce8-ba74-4368a26f5d44","Type":"ContainerStarted","Data":"d98d1003e94464d3ea2ae27c5ebb11913995cf8c0c5b029f9f8db5a79f4d35b2"} Feb 25 13:52:02 crc kubenswrapper[4815]: I0225 13:52:02.760161 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533792-b2jvf" podStartSLOduration=1.313548692 podStartE2EDuration="2.760141989s" podCreationTimestamp="2026-02-25 13:52:00 +0000 UTC" firstStartedPulling="2026-02-25 13:52:00.943457622 +0000 UTC m=+1878.744555676" lastFinishedPulling="2026-02-25 13:52:02.390050919 +0000 UTC m=+1880.191148973" observedRunningTime="2026-02-25 13:52:02.758332533 +0000 UTC m=+1880.559430617" watchObservedRunningTime="2026-02-25 13:52:02.760141989 +0000 UTC m=+1880.561240043" Feb 25 13:52:03 crc kubenswrapper[4815]: I0225 13:52:03.761108 4815 generic.go:334] "Generic (PLEG): container finished" podID="4449c14e-2d79-4ce8-ba74-4368a26f5d44" containerID="d98d1003e94464d3ea2ae27c5ebb11913995cf8c0c5b029f9f8db5a79f4d35b2" exitCode=0 Feb 25 13:52:03 crc kubenswrapper[4815]: I0225 13:52:03.761168 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533792-b2jvf" event={"ID":"4449c14e-2d79-4ce8-ba74-4368a26f5d44","Type":"ContainerDied","Data":"d98d1003e94464d3ea2ae27c5ebb11913995cf8c0c5b029f9f8db5a79f4d35b2"} Feb 25 13:52:05 crc kubenswrapper[4815]: I0225 13:52:05.178152 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533792-b2jvf" Feb 25 13:52:05 crc kubenswrapper[4815]: I0225 13:52:05.315732 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsfhn\" (UniqueName: \"kubernetes.io/projected/4449c14e-2d79-4ce8-ba74-4368a26f5d44-kube-api-access-nsfhn\") pod \"4449c14e-2d79-4ce8-ba74-4368a26f5d44\" (UID: \"4449c14e-2d79-4ce8-ba74-4368a26f5d44\") " Feb 25 13:52:05 crc kubenswrapper[4815]: I0225 13:52:05.324974 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4449c14e-2d79-4ce8-ba74-4368a26f5d44-kube-api-access-nsfhn" (OuterVolumeSpecName: "kube-api-access-nsfhn") pod "4449c14e-2d79-4ce8-ba74-4368a26f5d44" (UID: "4449c14e-2d79-4ce8-ba74-4368a26f5d44"). InnerVolumeSpecName "kube-api-access-nsfhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:52:05 crc kubenswrapper[4815]: I0225 13:52:05.417980 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsfhn\" (UniqueName: \"kubernetes.io/projected/4449c14e-2d79-4ce8-ba74-4368a26f5d44-kube-api-access-nsfhn\") on node \"crc\" DevicePath \"\"" Feb 25 13:52:05 crc kubenswrapper[4815]: I0225 13:52:05.783223 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533792-b2jvf" event={"ID":"4449c14e-2d79-4ce8-ba74-4368a26f5d44","Type":"ContainerDied","Data":"9f5c2836005f7d5f6293e71f661fce0ec3b5d1600821ca50bec91fe889e35e6a"} Feb 25 13:52:05 crc kubenswrapper[4815]: I0225 13:52:05.783266 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f5c2836005f7d5f6293e71f661fce0ec3b5d1600821ca50bec91fe889e35e6a" Feb 25 13:52:05 crc kubenswrapper[4815]: I0225 13:52:05.783331 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533792-b2jvf" Feb 25 13:52:05 crc kubenswrapper[4815]: I0225 13:52:05.836384 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533786-j5c6h"] Feb 25 13:52:05 crc kubenswrapper[4815]: I0225 13:52:05.843873 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533786-j5c6h"] Feb 25 13:52:06 crc kubenswrapper[4815]: I0225 13:52:06.949846 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="657b0b94-e86a-487a-98e6-54f445ccab68" path="/var/lib/kubelet/pods/657b0b94-e86a-487a-98e6-54f445ccab68/volumes" Feb 25 13:52:36 crc kubenswrapper[4815]: I0225 13:52:36.088631 4815 generic.go:334] "Generic (PLEG): container finished" podID="8d7f9d84-5806-4beb-9277-c77da4440527" containerID="f792b2707476dd0d63b9759613de232448f750d926051d29c127874a9dc2497c" exitCode=0 Feb 25 13:52:36 crc kubenswrapper[4815]: I0225 13:52:36.088748 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" event={"ID":"8d7f9d84-5806-4beb-9277-c77da4440527","Type":"ContainerDied","Data":"f792b2707476dd0d63b9759613de232448f750d926051d29c127874a9dc2497c"} Feb 25 13:52:37 crc kubenswrapper[4815]: I0225 13:52:37.496843 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:52:37 crc kubenswrapper[4815]: I0225 13:52:37.618665 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-inventory\") pod \"8d7f9d84-5806-4beb-9277-c77da4440527\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " Feb 25 13:52:37 crc kubenswrapper[4815]: I0225 13:52:37.619066 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-ssh-key-openstack-edpm-ipam\") pod \"8d7f9d84-5806-4beb-9277-c77da4440527\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " Feb 25 13:52:37 crc kubenswrapper[4815]: I0225 13:52:37.619192 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-ovn-combined-ca-bundle\") pod \"8d7f9d84-5806-4beb-9277-c77da4440527\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " Feb 25 13:52:37 crc kubenswrapper[4815]: I0225 13:52:37.619499 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8d7f9d84-5806-4beb-9277-c77da4440527-ovncontroller-config-0\") pod \"8d7f9d84-5806-4beb-9277-c77da4440527\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " Feb 25 13:52:37 crc kubenswrapper[4815]: I0225 13:52:37.619755 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tv6rk\" (UniqueName: \"kubernetes.io/projected/8d7f9d84-5806-4beb-9277-c77da4440527-kube-api-access-tv6rk\") pod \"8d7f9d84-5806-4beb-9277-c77da4440527\" (UID: \"8d7f9d84-5806-4beb-9277-c77da4440527\") " Feb 25 13:52:37 crc kubenswrapper[4815]: I0225 13:52:37.635948 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "8d7f9d84-5806-4beb-9277-c77da4440527" (UID: "8d7f9d84-5806-4beb-9277-c77da4440527"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:52:37 crc kubenswrapper[4815]: I0225 13:52:37.636222 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d7f9d84-5806-4beb-9277-c77da4440527-kube-api-access-tv6rk" (OuterVolumeSpecName: "kube-api-access-tv6rk") pod "8d7f9d84-5806-4beb-9277-c77da4440527" (UID: "8d7f9d84-5806-4beb-9277-c77da4440527"). InnerVolumeSpecName "kube-api-access-tv6rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:52:37 crc kubenswrapper[4815]: I0225 13:52:37.652107 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7f9d84-5806-4beb-9277-c77da4440527-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "8d7f9d84-5806-4beb-9277-c77da4440527" (UID: "8d7f9d84-5806-4beb-9277-c77da4440527"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:52:37 crc kubenswrapper[4815]: I0225 13:52:37.669730 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-inventory" (OuterVolumeSpecName: "inventory") pod "8d7f9d84-5806-4beb-9277-c77da4440527" (UID: "8d7f9d84-5806-4beb-9277-c77da4440527"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:52:37 crc kubenswrapper[4815]: I0225 13:52:37.674901 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8d7f9d84-5806-4beb-9277-c77da4440527" (UID: "8d7f9d84-5806-4beb-9277-c77da4440527"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:52:37 crc kubenswrapper[4815]: I0225 13:52:37.721900 4815 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:52:37 crc kubenswrapper[4815]: I0225 13:52:37.721944 4815 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8d7f9d84-5806-4beb-9277-c77da4440527-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:52:37 crc kubenswrapper[4815]: I0225 13:52:37.721955 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tv6rk\" (UniqueName: \"kubernetes.io/projected/8d7f9d84-5806-4beb-9277-c77da4440527-kube-api-access-tv6rk\") on node \"crc\" DevicePath \"\"" Feb 25 13:52:37 crc kubenswrapper[4815]: I0225 13:52:37.721963 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 13:52:37 crc kubenswrapper[4815]: I0225 13:52:37.721973 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d7f9d84-5806-4beb-9277-c77da4440527-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.112327 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" event={"ID":"8d7f9d84-5806-4beb-9277-c77da4440527","Type":"ContainerDied","Data":"1184df3214a105434046f8f7ae756f6702bbc6d4442f6d09a0fa1fb431b0c5ce"} Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.112395 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1184df3214a105434046f8f7ae756f6702bbc6d4442f6d09a0fa1fb431b0c5ce" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.112468 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jvq95" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.324269 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r"] Feb 25 13:52:38 crc kubenswrapper[4815]: E0225 13:52:38.324753 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4449c14e-2d79-4ce8-ba74-4368a26f5d44" containerName="oc" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.324778 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="4449c14e-2d79-4ce8-ba74-4368a26f5d44" containerName="oc" Feb 25 13:52:38 crc kubenswrapper[4815]: E0225 13:52:38.324803 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d7f9d84-5806-4beb-9277-c77da4440527" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.324813 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d7f9d84-5806-4beb-9277-c77da4440527" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.325043 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="4449c14e-2d79-4ce8-ba74-4368a26f5d44" containerName="oc" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.325070 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d7f9d84-5806-4beb-9277-c77da4440527" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.325839 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.328021 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.328307 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mmbj" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.328446 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.328583 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.328753 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.328789 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.342044 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r"] Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.437677 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.438046 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.438084 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.438335 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lnzp\" (UniqueName: \"kubernetes.io/projected/d2b331ae-52bd-4658-ae73-42674c94d3fa-kube-api-access-8lnzp\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.438479 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.438586 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.540835 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.541450 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.541577 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.541796 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lnzp\" (UniqueName: \"kubernetes.io/projected/d2b331ae-52bd-4658-ae73-42674c94d3fa-kube-api-access-8lnzp\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.541936 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.542054 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.546414 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.547229 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.547850 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.548095 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.554638 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.571979 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lnzp\" (UniqueName: \"kubernetes.io/projected/d2b331ae-52bd-4658-ae73-42674c94d3fa-kube-api-access-8lnzp\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:38 crc kubenswrapper[4815]: I0225 13:52:38.651548 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:52:39 crc kubenswrapper[4815]: I0225 13:52:39.183944 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r"] Feb 25 13:52:40 crc kubenswrapper[4815]: I0225 13:52:40.138420 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" event={"ID":"d2b331ae-52bd-4658-ae73-42674c94d3fa","Type":"ContainerStarted","Data":"c426adf0d8eb6f4f332e645f90aea356469563ebeb59f55853540d6e330f6445"} Feb 25 13:52:40 crc kubenswrapper[4815]: I0225 13:52:40.139064 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" event={"ID":"d2b331ae-52bd-4658-ae73-42674c94d3fa","Type":"ContainerStarted","Data":"bb02cfd11c139d564defd83087b5b9a1f1834fcb53562e6245c91a81011523ed"} Feb 25 13:52:48 crc kubenswrapper[4815]: I0225 13:52:48.132787 4815 scope.go:117] "RemoveContainer" containerID="c3ec72f599c5ebb4e95c902f08a9d31ba8859ee74bfe21b74301d6bd884f3850" Feb 25 13:53:25 crc kubenswrapper[4815]: I0225 13:53:25.644567 4815 generic.go:334] "Generic (PLEG): container finished" podID="d2b331ae-52bd-4658-ae73-42674c94d3fa" containerID="c426adf0d8eb6f4f332e645f90aea356469563ebeb59f55853540d6e330f6445" exitCode=0 Feb 25 13:53:25 crc kubenswrapper[4815]: I0225 13:53:25.644661 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" event={"ID":"d2b331ae-52bd-4658-ae73-42674c94d3fa","Type":"ContainerDied","Data":"c426adf0d8eb6f4f332e645f90aea356469563ebeb59f55853540d6e330f6445"} Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.057615 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.090035 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-neutron-metadata-combined-ca-bundle\") pod \"d2b331ae-52bd-4658-ae73-42674c94d3fa\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.090236 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-neutron-ovn-metadata-agent-neutron-config-0\") pod \"d2b331ae-52bd-4658-ae73-42674c94d3fa\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.091216 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-inventory\") pod \"d2b331ae-52bd-4658-ae73-42674c94d3fa\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.091283 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-ssh-key-openstack-edpm-ipam\") pod \"d2b331ae-52bd-4658-ae73-42674c94d3fa\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.091695 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lnzp\" (UniqueName: \"kubernetes.io/projected/d2b331ae-52bd-4658-ae73-42674c94d3fa-kube-api-access-8lnzp\") pod \"d2b331ae-52bd-4658-ae73-42674c94d3fa\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.091838 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-nova-metadata-neutron-config-0\") pod \"d2b331ae-52bd-4658-ae73-42674c94d3fa\" (UID: \"d2b331ae-52bd-4658-ae73-42674c94d3fa\") " Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.096942 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d2b331ae-52bd-4658-ae73-42674c94d3fa" (UID: "d2b331ae-52bd-4658-ae73-42674c94d3fa"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.099620 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2b331ae-52bd-4658-ae73-42674c94d3fa-kube-api-access-8lnzp" (OuterVolumeSpecName: "kube-api-access-8lnzp") pod "d2b331ae-52bd-4658-ae73-42674c94d3fa" (UID: "d2b331ae-52bd-4658-ae73-42674c94d3fa"). InnerVolumeSpecName "kube-api-access-8lnzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.123047 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "d2b331ae-52bd-4658-ae73-42674c94d3fa" (UID: "d2b331ae-52bd-4658-ae73-42674c94d3fa"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.130133 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d2b331ae-52bd-4658-ae73-42674c94d3fa" (UID: "d2b331ae-52bd-4658-ae73-42674c94d3fa"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.131358 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "d2b331ae-52bd-4658-ae73-42674c94d3fa" (UID: "d2b331ae-52bd-4658-ae73-42674c94d3fa"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.147946 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-inventory" (OuterVolumeSpecName: "inventory") pod "d2b331ae-52bd-4658-ae73-42674c94d3fa" (UID: "d2b331ae-52bd-4658-ae73-42674c94d3fa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.194361 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.194887 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.195147 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lnzp\" (UniqueName: \"kubernetes.io/projected/d2b331ae-52bd-4658-ae73-42674c94d3fa-kube-api-access-8lnzp\") on node \"crc\" DevicePath \"\"" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.195185 4815 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.195199 4815 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.195212 4815 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d2b331ae-52bd-4658-ae73-42674c94d3fa-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.665343 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" event={"ID":"d2b331ae-52bd-4658-ae73-42674c94d3fa","Type":"ContainerDied","Data":"bb02cfd11c139d564defd83087b5b9a1f1834fcb53562e6245c91a81011523ed"} Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.665664 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb02cfd11c139d564defd83087b5b9a1f1834fcb53562e6245c91a81011523ed" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.665492 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.764036 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm"] Feb 25 13:53:27 crc kubenswrapper[4815]: E0225 13:53:27.764540 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2b331ae-52bd-4658-ae73-42674c94d3fa" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.764567 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2b331ae-52bd-4658-ae73-42674c94d3fa" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.764831 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2b331ae-52bd-4658-ae73-42674c94d3fa" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.765698 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.768413 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.768774 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mmbj" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.768808 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.769058 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.769328 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.782360 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm"] Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.812684 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrwgb\" (UniqueName: \"kubernetes.io/projected/4e6edebf-da64-4665-9781-b4f3191e0d10-kube-api-access-mrwgb\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-85zsm\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.812742 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-85zsm\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.812992 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-85zsm\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.813248 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-85zsm\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.813375 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-85zsm\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.915476 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-85zsm\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.915594 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrwgb\" (UniqueName: \"kubernetes.io/projected/4e6edebf-da64-4665-9781-b4f3191e0d10-kube-api-access-mrwgb\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-85zsm\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.915659 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-85zsm\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.915806 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-85zsm\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.915899 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-85zsm\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.922214 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-85zsm\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.922636 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-85zsm\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.923123 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-85zsm\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.924781 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-85zsm\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:53:27 crc kubenswrapper[4815]: I0225 13:53:27.942692 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrwgb\" (UniqueName: \"kubernetes.io/projected/4e6edebf-da64-4665-9781-b4f3191e0d10-kube-api-access-mrwgb\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-85zsm\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:53:28 crc kubenswrapper[4815]: I0225 13:53:28.083279 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:53:28 crc kubenswrapper[4815]: I0225 13:53:28.655909 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm"] Feb 25 13:53:28 crc kubenswrapper[4815]: I0225 13:53:28.677870 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" event={"ID":"4e6edebf-da64-4665-9781-b4f3191e0d10","Type":"ContainerStarted","Data":"07ca3b66ec0c4ae425a50946b5e980c3f4804b97e2092340685bcbc9468dd4b0"} Feb 25 13:53:29 crc kubenswrapper[4815]: I0225 13:53:29.689216 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" event={"ID":"4e6edebf-da64-4665-9781-b4f3191e0d10","Type":"ContainerStarted","Data":"ffbed81d33216662483df14d093ab90e9bfc9726d21a12a22093bc0fdaffbb08"} Feb 25 13:53:29 crc kubenswrapper[4815]: I0225 13:53:29.714573 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" podStartSLOduration=2.2747705959999998 podStartE2EDuration="2.714554857s" podCreationTimestamp="2026-02-25 13:53:27 +0000 UTC" firstStartedPulling="2026-02-25 13:53:28.664192972 +0000 UTC m=+1966.465291026" lastFinishedPulling="2026-02-25 13:53:29.103977193 +0000 UTC m=+1966.905075287" observedRunningTime="2026-02-25 13:53:29.709263552 +0000 UTC m=+1967.510361666" watchObservedRunningTime="2026-02-25 13:53:29.714554857 +0000 UTC m=+1967.515652911" Feb 25 13:53:51 crc kubenswrapper[4815]: I0225 13:53:51.358050 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:53:51 crc kubenswrapper[4815]: I0225 13:53:51.358729 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:54:00 crc kubenswrapper[4815]: I0225 13:54:00.157397 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533794-l8z6b"] Feb 25 13:54:00 crc kubenswrapper[4815]: I0225 13:54:00.159419 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533794-l8z6b" Feb 25 13:54:00 crc kubenswrapper[4815]: I0225 13:54:00.162108 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:54:00 crc kubenswrapper[4815]: I0225 13:54:00.162116 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:54:00 crc kubenswrapper[4815]: I0225 13:54:00.162131 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:54:00 crc kubenswrapper[4815]: I0225 13:54:00.166850 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533794-l8z6b"] Feb 25 13:54:00 crc kubenswrapper[4815]: I0225 13:54:00.258868 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m6cx\" (UniqueName: \"kubernetes.io/projected/fc77125d-9f98-41b8-a366-d35124c45f9f-kube-api-access-4m6cx\") pod \"auto-csr-approver-29533794-l8z6b\" (UID: \"fc77125d-9f98-41b8-a366-d35124c45f9f\") " pod="openshift-infra/auto-csr-approver-29533794-l8z6b" Feb 25 13:54:00 crc kubenswrapper[4815]: I0225 13:54:00.362183 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m6cx\" (UniqueName: \"kubernetes.io/projected/fc77125d-9f98-41b8-a366-d35124c45f9f-kube-api-access-4m6cx\") pod \"auto-csr-approver-29533794-l8z6b\" (UID: \"fc77125d-9f98-41b8-a366-d35124c45f9f\") " pod="openshift-infra/auto-csr-approver-29533794-l8z6b" Feb 25 13:54:00 crc kubenswrapper[4815]: I0225 13:54:00.394553 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m6cx\" (UniqueName: \"kubernetes.io/projected/fc77125d-9f98-41b8-a366-d35124c45f9f-kube-api-access-4m6cx\") pod \"auto-csr-approver-29533794-l8z6b\" (UID: \"fc77125d-9f98-41b8-a366-d35124c45f9f\") " pod="openshift-infra/auto-csr-approver-29533794-l8z6b" Feb 25 13:54:00 crc kubenswrapper[4815]: I0225 13:54:00.498136 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533794-l8z6b" Feb 25 13:54:01 crc kubenswrapper[4815]: I0225 13:54:01.072116 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533794-l8z6b"] Feb 25 13:54:02 crc kubenswrapper[4815]: I0225 13:54:02.000601 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533794-l8z6b" event={"ID":"fc77125d-9f98-41b8-a366-d35124c45f9f","Type":"ContainerStarted","Data":"8fcdf0d4a04ec4d63aecab654f9ca5245e5fb76667023d3e8700469c54c8d0c6"} Feb 25 13:54:03 crc kubenswrapper[4815]: I0225 13:54:03.013240 4815 generic.go:334] "Generic (PLEG): container finished" podID="fc77125d-9f98-41b8-a366-d35124c45f9f" containerID="ea3d582b548bf242a266e784b53beb6b69a0e08bfae98fb3536a7c6fed44fe63" exitCode=0 Feb 25 13:54:03 crc kubenswrapper[4815]: I0225 13:54:03.013352 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533794-l8z6b" event={"ID":"fc77125d-9f98-41b8-a366-d35124c45f9f","Type":"ContainerDied","Data":"ea3d582b548bf242a266e784b53beb6b69a0e08bfae98fb3536a7c6fed44fe63"} Feb 25 13:54:04 crc kubenswrapper[4815]: I0225 13:54:04.357836 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533794-l8z6b" Feb 25 13:54:04 crc kubenswrapper[4815]: I0225 13:54:04.462067 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4m6cx\" (UniqueName: \"kubernetes.io/projected/fc77125d-9f98-41b8-a366-d35124c45f9f-kube-api-access-4m6cx\") pod \"fc77125d-9f98-41b8-a366-d35124c45f9f\" (UID: \"fc77125d-9f98-41b8-a366-d35124c45f9f\") " Feb 25 13:54:04 crc kubenswrapper[4815]: I0225 13:54:04.467914 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc77125d-9f98-41b8-a366-d35124c45f9f-kube-api-access-4m6cx" (OuterVolumeSpecName: "kube-api-access-4m6cx") pod "fc77125d-9f98-41b8-a366-d35124c45f9f" (UID: "fc77125d-9f98-41b8-a366-d35124c45f9f"). InnerVolumeSpecName "kube-api-access-4m6cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:54:04 crc kubenswrapper[4815]: I0225 13:54:04.565005 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4m6cx\" (UniqueName: \"kubernetes.io/projected/fc77125d-9f98-41b8-a366-d35124c45f9f-kube-api-access-4m6cx\") on node \"crc\" DevicePath \"\"" Feb 25 13:54:05 crc kubenswrapper[4815]: I0225 13:54:05.031127 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533794-l8z6b" event={"ID":"fc77125d-9f98-41b8-a366-d35124c45f9f","Type":"ContainerDied","Data":"8fcdf0d4a04ec4d63aecab654f9ca5245e5fb76667023d3e8700469c54c8d0c6"} Feb 25 13:54:05 crc kubenswrapper[4815]: I0225 13:54:05.031165 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fcdf0d4a04ec4d63aecab654f9ca5245e5fb76667023d3e8700469c54c8d0c6" Feb 25 13:54:05 crc kubenswrapper[4815]: I0225 13:54:05.031165 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533794-l8z6b" Feb 25 13:54:05 crc kubenswrapper[4815]: I0225 13:54:05.440628 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533788-7kz4t"] Feb 25 13:54:05 crc kubenswrapper[4815]: I0225 13:54:05.449887 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533788-7kz4t"] Feb 25 13:54:06 crc kubenswrapper[4815]: I0225 13:54:06.947698 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9533594f-c28a-46a8-b87b-2d872b53d1b3" path="/var/lib/kubelet/pods/9533594f-c28a-46a8-b87b-2d872b53d1b3/volumes" Feb 25 13:54:21 crc kubenswrapper[4815]: I0225 13:54:21.342381 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:54:21 crc kubenswrapper[4815]: I0225 13:54:21.342876 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:54:48 crc kubenswrapper[4815]: I0225 13:54:48.237396 4815 scope.go:117] "RemoveContainer" containerID="827676a0a42a6d72a7625fbd961d1c9d4cba3ce0e673f3a31d2f446c05118cd9" Feb 25 13:54:51 crc kubenswrapper[4815]: I0225 13:54:51.341861 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:54:51 crc kubenswrapper[4815]: I0225 13:54:51.342229 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:54:51 crc kubenswrapper[4815]: I0225 13:54:51.342286 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:54:51 crc kubenswrapper[4815]: I0225 13:54:51.342995 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bc1a87ba95ff2fe57f3edfcdf517b6e2b034a474d761b615e638c4a52c55574e"} pod="openshift-machine-config-operator/machine-config-daemon-frc75" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 13:54:51 crc kubenswrapper[4815]: I0225 13:54:51.343068 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" containerID="cri-o://bc1a87ba95ff2fe57f3edfcdf517b6e2b034a474d761b615e638c4a52c55574e" gracePeriod=600 Feb 25 13:54:51 crc kubenswrapper[4815]: I0225 13:54:51.481058 4815 generic.go:334] "Generic (PLEG): container finished" podID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerID="bc1a87ba95ff2fe57f3edfcdf517b6e2b034a474d761b615e638c4a52c55574e" exitCode=0 Feb 25 13:54:51 crc kubenswrapper[4815]: I0225 13:54:51.481150 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerDied","Data":"bc1a87ba95ff2fe57f3edfcdf517b6e2b034a474d761b615e638c4a52c55574e"} Feb 25 13:54:51 crc kubenswrapper[4815]: I0225 13:54:51.481410 4815 scope.go:117] "RemoveContainer" containerID="faf23760e4c2b1d88c895ce40ee614ba9c2a710dfe3bb6bd7e7ed425d241a10b" Feb 25 13:54:52 crc kubenswrapper[4815]: I0225 13:54:52.492412 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerStarted","Data":"9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05"} Feb 25 13:55:21 crc kubenswrapper[4815]: I0225 13:55:21.043844 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7b5sq"] Feb 25 13:55:21 crc kubenswrapper[4815]: E0225 13:55:21.044797 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc77125d-9f98-41b8-a366-d35124c45f9f" containerName="oc" Feb 25 13:55:21 crc kubenswrapper[4815]: I0225 13:55:21.044812 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc77125d-9f98-41b8-a366-d35124c45f9f" containerName="oc" Feb 25 13:55:21 crc kubenswrapper[4815]: I0225 13:55:21.045066 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc77125d-9f98-41b8-a366-d35124c45f9f" containerName="oc" Feb 25 13:55:21 crc kubenswrapper[4815]: I0225 13:55:21.046894 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7b5sq" Feb 25 13:55:21 crc kubenswrapper[4815]: I0225 13:55:21.055611 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7b5sq"] Feb 25 13:55:21 crc kubenswrapper[4815]: I0225 13:55:21.130454 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a940fb1-ad85-47bf-be31-5a607d684e3c-catalog-content\") pod \"redhat-marketplace-7b5sq\" (UID: \"1a940fb1-ad85-47bf-be31-5a607d684e3c\") " pod="openshift-marketplace/redhat-marketplace-7b5sq" Feb 25 13:55:21 crc kubenswrapper[4815]: I0225 13:55:21.130957 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a940fb1-ad85-47bf-be31-5a607d684e3c-utilities\") pod \"redhat-marketplace-7b5sq\" (UID: \"1a940fb1-ad85-47bf-be31-5a607d684e3c\") " pod="openshift-marketplace/redhat-marketplace-7b5sq" Feb 25 13:55:21 crc kubenswrapper[4815]: I0225 13:55:21.131113 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xghdl\" (UniqueName: \"kubernetes.io/projected/1a940fb1-ad85-47bf-be31-5a607d684e3c-kube-api-access-xghdl\") pod \"redhat-marketplace-7b5sq\" (UID: \"1a940fb1-ad85-47bf-be31-5a607d684e3c\") " pod="openshift-marketplace/redhat-marketplace-7b5sq" Feb 25 13:55:21 crc kubenswrapper[4815]: I0225 13:55:21.233247 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a940fb1-ad85-47bf-be31-5a607d684e3c-utilities\") pod \"redhat-marketplace-7b5sq\" (UID: \"1a940fb1-ad85-47bf-be31-5a607d684e3c\") " pod="openshift-marketplace/redhat-marketplace-7b5sq" Feb 25 13:55:21 crc kubenswrapper[4815]: I0225 13:55:21.233305 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xghdl\" (UniqueName: \"kubernetes.io/projected/1a940fb1-ad85-47bf-be31-5a607d684e3c-kube-api-access-xghdl\") pod \"redhat-marketplace-7b5sq\" (UID: \"1a940fb1-ad85-47bf-be31-5a607d684e3c\") " pod="openshift-marketplace/redhat-marketplace-7b5sq" Feb 25 13:55:21 crc kubenswrapper[4815]: I0225 13:55:21.233352 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a940fb1-ad85-47bf-be31-5a607d684e3c-catalog-content\") pod \"redhat-marketplace-7b5sq\" (UID: \"1a940fb1-ad85-47bf-be31-5a607d684e3c\") " pod="openshift-marketplace/redhat-marketplace-7b5sq" Feb 25 13:55:21 crc kubenswrapper[4815]: I0225 13:55:21.233874 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a940fb1-ad85-47bf-be31-5a607d684e3c-utilities\") pod \"redhat-marketplace-7b5sq\" (UID: \"1a940fb1-ad85-47bf-be31-5a607d684e3c\") " pod="openshift-marketplace/redhat-marketplace-7b5sq" Feb 25 13:55:21 crc kubenswrapper[4815]: I0225 13:55:21.233921 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a940fb1-ad85-47bf-be31-5a607d684e3c-catalog-content\") pod \"redhat-marketplace-7b5sq\" (UID: \"1a940fb1-ad85-47bf-be31-5a607d684e3c\") " pod="openshift-marketplace/redhat-marketplace-7b5sq" Feb 25 13:55:21 crc kubenswrapper[4815]: I0225 13:55:21.258578 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xghdl\" (UniqueName: \"kubernetes.io/projected/1a940fb1-ad85-47bf-be31-5a607d684e3c-kube-api-access-xghdl\") pod \"redhat-marketplace-7b5sq\" (UID: \"1a940fb1-ad85-47bf-be31-5a607d684e3c\") " pod="openshift-marketplace/redhat-marketplace-7b5sq" Feb 25 13:55:21 crc kubenswrapper[4815]: I0225 13:55:21.371378 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7b5sq" Feb 25 13:55:21 crc kubenswrapper[4815]: I0225 13:55:21.882914 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7b5sq"] Feb 25 13:55:22 crc kubenswrapper[4815]: I0225 13:55:22.802207 4815 generic.go:334] "Generic (PLEG): container finished" podID="1a940fb1-ad85-47bf-be31-5a607d684e3c" containerID="76177892ab39ed91c4c37af1448a6eb3de80ac5f671e9c86ca9a1316644f98e4" exitCode=0 Feb 25 13:55:22 crc kubenswrapper[4815]: I0225 13:55:22.802548 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7b5sq" event={"ID":"1a940fb1-ad85-47bf-be31-5a607d684e3c","Type":"ContainerDied","Data":"76177892ab39ed91c4c37af1448a6eb3de80ac5f671e9c86ca9a1316644f98e4"} Feb 25 13:55:22 crc kubenswrapper[4815]: I0225 13:55:22.802582 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7b5sq" event={"ID":"1a940fb1-ad85-47bf-be31-5a607d684e3c","Type":"ContainerStarted","Data":"ef97e9ccefdefa97cf34c5c2f111737d49eeeb36a7368fe726ab95a215058c08"} Feb 25 13:55:22 crc kubenswrapper[4815]: I0225 13:55:22.805874 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 13:55:23 crc kubenswrapper[4815]: I0225 13:55:23.814977 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7b5sq" event={"ID":"1a940fb1-ad85-47bf-be31-5a607d684e3c","Type":"ContainerStarted","Data":"d10c67de76f5a00d0b8331a2cd3d0766a8afafe8b102822ffdfa982d2c2cd30d"} Feb 25 13:55:24 crc kubenswrapper[4815]: I0225 13:55:24.828633 4815 generic.go:334] "Generic (PLEG): container finished" podID="1a940fb1-ad85-47bf-be31-5a607d684e3c" containerID="d10c67de76f5a00d0b8331a2cd3d0766a8afafe8b102822ffdfa982d2c2cd30d" exitCode=0 Feb 25 13:55:24 crc kubenswrapper[4815]: I0225 13:55:24.828759 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7b5sq" event={"ID":"1a940fb1-ad85-47bf-be31-5a607d684e3c","Type":"ContainerDied","Data":"d10c67de76f5a00d0b8331a2cd3d0766a8afafe8b102822ffdfa982d2c2cd30d"} Feb 25 13:55:25 crc kubenswrapper[4815]: I0225 13:55:25.843176 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7b5sq" event={"ID":"1a940fb1-ad85-47bf-be31-5a607d684e3c","Type":"ContainerStarted","Data":"04a3b9fcafe50e0c255ec8a1efdea8c5e4571a5411aaf6d3bf963f084e5334bc"} Feb 25 13:55:25 crc kubenswrapper[4815]: I0225 13:55:25.885024 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7b5sq" podStartSLOduration=2.447245617 podStartE2EDuration="4.884996526s" podCreationTimestamp="2026-02-25 13:55:21 +0000 UTC" firstStartedPulling="2026-02-25 13:55:22.805145618 +0000 UTC m=+2080.606243712" lastFinishedPulling="2026-02-25 13:55:25.242896567 +0000 UTC m=+2083.043994621" observedRunningTime="2026-02-25 13:55:25.860124629 +0000 UTC m=+2083.661222703" watchObservedRunningTime="2026-02-25 13:55:25.884996526 +0000 UTC m=+2083.686094590" Feb 25 13:55:31 crc kubenswrapper[4815]: I0225 13:55:31.371831 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7b5sq" Feb 25 13:55:31 crc kubenswrapper[4815]: I0225 13:55:31.372983 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7b5sq" Feb 25 13:55:31 crc kubenswrapper[4815]: I0225 13:55:31.427623 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7b5sq" Feb 25 13:55:31 crc kubenswrapper[4815]: I0225 13:55:31.953118 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7b5sq" Feb 25 13:55:33 crc kubenswrapper[4815]: I0225 13:55:33.022883 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7b5sq"] Feb 25 13:55:33 crc kubenswrapper[4815]: I0225 13:55:33.924095 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7b5sq" podUID="1a940fb1-ad85-47bf-be31-5a607d684e3c" containerName="registry-server" containerID="cri-o://04a3b9fcafe50e0c255ec8a1efdea8c5e4571a5411aaf6d3bf963f084e5334bc" gracePeriod=2 Feb 25 13:55:34 crc kubenswrapper[4815]: I0225 13:55:34.415172 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7b5sq" Feb 25 13:55:34 crc kubenswrapper[4815]: I0225 13:55:34.587095 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a940fb1-ad85-47bf-be31-5a607d684e3c-utilities\") pod \"1a940fb1-ad85-47bf-be31-5a607d684e3c\" (UID: \"1a940fb1-ad85-47bf-be31-5a607d684e3c\") " Feb 25 13:55:34 crc kubenswrapper[4815]: I0225 13:55:34.587166 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a940fb1-ad85-47bf-be31-5a607d684e3c-catalog-content\") pod \"1a940fb1-ad85-47bf-be31-5a607d684e3c\" (UID: \"1a940fb1-ad85-47bf-be31-5a607d684e3c\") " Feb 25 13:55:34 crc kubenswrapper[4815]: I0225 13:55:34.587354 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xghdl\" (UniqueName: \"kubernetes.io/projected/1a940fb1-ad85-47bf-be31-5a607d684e3c-kube-api-access-xghdl\") pod \"1a940fb1-ad85-47bf-be31-5a607d684e3c\" (UID: \"1a940fb1-ad85-47bf-be31-5a607d684e3c\") " Feb 25 13:55:34 crc kubenswrapper[4815]: I0225 13:55:34.588321 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a940fb1-ad85-47bf-be31-5a607d684e3c-utilities" (OuterVolumeSpecName: "utilities") pod "1a940fb1-ad85-47bf-be31-5a607d684e3c" (UID: "1a940fb1-ad85-47bf-be31-5a607d684e3c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:55:34 crc kubenswrapper[4815]: I0225 13:55:34.601373 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a940fb1-ad85-47bf-be31-5a607d684e3c-kube-api-access-xghdl" (OuterVolumeSpecName: "kube-api-access-xghdl") pod "1a940fb1-ad85-47bf-be31-5a607d684e3c" (UID: "1a940fb1-ad85-47bf-be31-5a607d684e3c"). InnerVolumeSpecName "kube-api-access-xghdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:55:34 crc kubenswrapper[4815]: I0225 13:55:34.619248 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a940fb1-ad85-47bf-be31-5a607d684e3c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a940fb1-ad85-47bf-be31-5a607d684e3c" (UID: "1a940fb1-ad85-47bf-be31-5a607d684e3c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:55:34 crc kubenswrapper[4815]: I0225 13:55:34.689194 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xghdl\" (UniqueName: \"kubernetes.io/projected/1a940fb1-ad85-47bf-be31-5a607d684e3c-kube-api-access-xghdl\") on node \"crc\" DevicePath \"\"" Feb 25 13:55:34 crc kubenswrapper[4815]: I0225 13:55:34.689230 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a940fb1-ad85-47bf-be31-5a607d684e3c-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:55:34 crc kubenswrapper[4815]: I0225 13:55:34.689241 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a940fb1-ad85-47bf-be31-5a607d684e3c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:55:34 crc kubenswrapper[4815]: I0225 13:55:34.949072 4815 generic.go:334] "Generic (PLEG): container finished" podID="1a940fb1-ad85-47bf-be31-5a607d684e3c" containerID="04a3b9fcafe50e0c255ec8a1efdea8c5e4571a5411aaf6d3bf963f084e5334bc" exitCode=0 Feb 25 13:55:34 crc kubenswrapper[4815]: I0225 13:55:34.949191 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7b5sq" Feb 25 13:55:34 crc kubenswrapper[4815]: I0225 13:55:34.956098 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7b5sq" event={"ID":"1a940fb1-ad85-47bf-be31-5a607d684e3c","Type":"ContainerDied","Data":"04a3b9fcafe50e0c255ec8a1efdea8c5e4571a5411aaf6d3bf963f084e5334bc"} Feb 25 13:55:34 crc kubenswrapper[4815]: I0225 13:55:34.956444 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7b5sq" event={"ID":"1a940fb1-ad85-47bf-be31-5a607d684e3c","Type":"ContainerDied","Data":"ef97e9ccefdefa97cf34c5c2f111737d49eeeb36a7368fe726ab95a215058c08"} Feb 25 13:55:34 crc kubenswrapper[4815]: I0225 13:55:34.956470 4815 scope.go:117] "RemoveContainer" containerID="04a3b9fcafe50e0c255ec8a1efdea8c5e4571a5411aaf6d3bf963f084e5334bc" Feb 25 13:55:34 crc kubenswrapper[4815]: I0225 13:55:34.989248 4815 scope.go:117] "RemoveContainer" containerID="d10c67de76f5a00d0b8331a2cd3d0766a8afafe8b102822ffdfa982d2c2cd30d" Feb 25 13:55:34 crc kubenswrapper[4815]: I0225 13:55:34.991458 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7b5sq"] Feb 25 13:55:35 crc kubenswrapper[4815]: I0225 13:55:34.998654 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7b5sq"] Feb 25 13:55:35 crc kubenswrapper[4815]: I0225 13:55:35.017105 4815 scope.go:117] "RemoveContainer" containerID="76177892ab39ed91c4c37af1448a6eb3de80ac5f671e9c86ca9a1316644f98e4" Feb 25 13:55:35 crc kubenswrapper[4815]: I0225 13:55:35.064707 4815 scope.go:117] "RemoveContainer" containerID="04a3b9fcafe50e0c255ec8a1efdea8c5e4571a5411aaf6d3bf963f084e5334bc" Feb 25 13:55:35 crc kubenswrapper[4815]: E0225 13:55:35.065648 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04a3b9fcafe50e0c255ec8a1efdea8c5e4571a5411aaf6d3bf963f084e5334bc\": container with ID starting with 04a3b9fcafe50e0c255ec8a1efdea8c5e4571a5411aaf6d3bf963f084e5334bc not found: ID does not exist" containerID="04a3b9fcafe50e0c255ec8a1efdea8c5e4571a5411aaf6d3bf963f084e5334bc" Feb 25 13:55:35 crc kubenswrapper[4815]: I0225 13:55:35.065913 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04a3b9fcafe50e0c255ec8a1efdea8c5e4571a5411aaf6d3bf963f084e5334bc"} err="failed to get container status \"04a3b9fcafe50e0c255ec8a1efdea8c5e4571a5411aaf6d3bf963f084e5334bc\": rpc error: code = NotFound desc = could not find container \"04a3b9fcafe50e0c255ec8a1efdea8c5e4571a5411aaf6d3bf963f084e5334bc\": container with ID starting with 04a3b9fcafe50e0c255ec8a1efdea8c5e4571a5411aaf6d3bf963f084e5334bc not found: ID does not exist" Feb 25 13:55:35 crc kubenswrapper[4815]: I0225 13:55:35.065956 4815 scope.go:117] "RemoveContainer" containerID="d10c67de76f5a00d0b8331a2cd3d0766a8afafe8b102822ffdfa982d2c2cd30d" Feb 25 13:55:35 crc kubenswrapper[4815]: E0225 13:55:35.066365 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d10c67de76f5a00d0b8331a2cd3d0766a8afafe8b102822ffdfa982d2c2cd30d\": container with ID starting with d10c67de76f5a00d0b8331a2cd3d0766a8afafe8b102822ffdfa982d2c2cd30d not found: ID does not exist" containerID="d10c67de76f5a00d0b8331a2cd3d0766a8afafe8b102822ffdfa982d2c2cd30d" Feb 25 13:55:35 crc kubenswrapper[4815]: I0225 13:55:35.066396 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d10c67de76f5a00d0b8331a2cd3d0766a8afafe8b102822ffdfa982d2c2cd30d"} err="failed to get container status \"d10c67de76f5a00d0b8331a2cd3d0766a8afafe8b102822ffdfa982d2c2cd30d\": rpc error: code = NotFound desc = could not find container \"d10c67de76f5a00d0b8331a2cd3d0766a8afafe8b102822ffdfa982d2c2cd30d\": container with ID starting with d10c67de76f5a00d0b8331a2cd3d0766a8afafe8b102822ffdfa982d2c2cd30d not found: ID does not exist" Feb 25 13:55:35 crc kubenswrapper[4815]: I0225 13:55:35.066447 4815 scope.go:117] "RemoveContainer" containerID="76177892ab39ed91c4c37af1448a6eb3de80ac5f671e9c86ca9a1316644f98e4" Feb 25 13:55:35 crc kubenswrapper[4815]: E0225 13:55:35.066876 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76177892ab39ed91c4c37af1448a6eb3de80ac5f671e9c86ca9a1316644f98e4\": container with ID starting with 76177892ab39ed91c4c37af1448a6eb3de80ac5f671e9c86ca9a1316644f98e4 not found: ID does not exist" containerID="76177892ab39ed91c4c37af1448a6eb3de80ac5f671e9c86ca9a1316644f98e4" Feb 25 13:55:35 crc kubenswrapper[4815]: I0225 13:55:35.066919 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76177892ab39ed91c4c37af1448a6eb3de80ac5f671e9c86ca9a1316644f98e4"} err="failed to get container status \"76177892ab39ed91c4c37af1448a6eb3de80ac5f671e9c86ca9a1316644f98e4\": rpc error: code = NotFound desc = could not find container \"76177892ab39ed91c4c37af1448a6eb3de80ac5f671e9c86ca9a1316644f98e4\": container with ID starting with 76177892ab39ed91c4c37af1448a6eb3de80ac5f671e9c86ca9a1316644f98e4 not found: ID does not exist" Feb 25 13:55:36 crc kubenswrapper[4815]: I0225 13:55:36.958075 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a940fb1-ad85-47bf-be31-5a607d684e3c" path="/var/lib/kubelet/pods/1a940fb1-ad85-47bf-be31-5a607d684e3c/volumes" Feb 25 13:56:00 crc kubenswrapper[4815]: I0225 13:56:00.143264 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533796-68pz9"] Feb 25 13:56:00 crc kubenswrapper[4815]: E0225 13:56:00.144396 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a940fb1-ad85-47bf-be31-5a607d684e3c" containerName="extract-content" Feb 25 13:56:00 crc kubenswrapper[4815]: I0225 13:56:00.144412 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a940fb1-ad85-47bf-be31-5a607d684e3c" containerName="extract-content" Feb 25 13:56:00 crc kubenswrapper[4815]: E0225 13:56:00.144439 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a940fb1-ad85-47bf-be31-5a607d684e3c" containerName="extract-utilities" Feb 25 13:56:00 crc kubenswrapper[4815]: I0225 13:56:00.144449 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a940fb1-ad85-47bf-be31-5a607d684e3c" containerName="extract-utilities" Feb 25 13:56:00 crc kubenswrapper[4815]: E0225 13:56:00.144469 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a940fb1-ad85-47bf-be31-5a607d684e3c" containerName="registry-server" Feb 25 13:56:00 crc kubenswrapper[4815]: I0225 13:56:00.144478 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a940fb1-ad85-47bf-be31-5a607d684e3c" containerName="registry-server" Feb 25 13:56:00 crc kubenswrapper[4815]: I0225 13:56:00.144757 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a940fb1-ad85-47bf-be31-5a607d684e3c" containerName="registry-server" Feb 25 13:56:00 crc kubenswrapper[4815]: I0225 13:56:00.145589 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533796-68pz9" Feb 25 13:56:00 crc kubenswrapper[4815]: I0225 13:56:00.148063 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:56:00 crc kubenswrapper[4815]: I0225 13:56:00.148134 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:56:00 crc kubenswrapper[4815]: I0225 13:56:00.148828 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:56:00 crc kubenswrapper[4815]: I0225 13:56:00.159753 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533796-68pz9"] Feb 25 13:56:00 crc kubenswrapper[4815]: I0225 13:56:00.274013 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x82kl\" (UniqueName: \"kubernetes.io/projected/7d5d295f-9999-4500-af75-c46724b26d89-kube-api-access-x82kl\") pod \"auto-csr-approver-29533796-68pz9\" (UID: \"7d5d295f-9999-4500-af75-c46724b26d89\") " pod="openshift-infra/auto-csr-approver-29533796-68pz9" Feb 25 13:56:00 crc kubenswrapper[4815]: I0225 13:56:00.376429 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x82kl\" (UniqueName: \"kubernetes.io/projected/7d5d295f-9999-4500-af75-c46724b26d89-kube-api-access-x82kl\") pod \"auto-csr-approver-29533796-68pz9\" (UID: \"7d5d295f-9999-4500-af75-c46724b26d89\") " pod="openshift-infra/auto-csr-approver-29533796-68pz9" Feb 25 13:56:00 crc kubenswrapper[4815]: I0225 13:56:00.410106 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x82kl\" (UniqueName: \"kubernetes.io/projected/7d5d295f-9999-4500-af75-c46724b26d89-kube-api-access-x82kl\") pod \"auto-csr-approver-29533796-68pz9\" (UID: \"7d5d295f-9999-4500-af75-c46724b26d89\") " pod="openshift-infra/auto-csr-approver-29533796-68pz9" Feb 25 13:56:00 crc kubenswrapper[4815]: I0225 13:56:00.473080 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533796-68pz9" Feb 25 13:56:00 crc kubenswrapper[4815]: I0225 13:56:00.967641 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533796-68pz9"] Feb 25 13:56:00 crc kubenswrapper[4815]: W0225 13:56:00.970815 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d5d295f_9999_4500_af75_c46724b26d89.slice/crio-2d3e3870cc305292e4febac36f113f94c247b6b511226e13b95bbd5cc90a531b WatchSource:0}: Error finding container 2d3e3870cc305292e4febac36f113f94c247b6b511226e13b95bbd5cc90a531b: Status 404 returned error can't find the container with id 2d3e3870cc305292e4febac36f113f94c247b6b511226e13b95bbd5cc90a531b Feb 25 13:56:01 crc kubenswrapper[4815]: I0225 13:56:01.267246 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533796-68pz9" event={"ID":"7d5d295f-9999-4500-af75-c46724b26d89","Type":"ContainerStarted","Data":"2d3e3870cc305292e4febac36f113f94c247b6b511226e13b95bbd5cc90a531b"} Feb 25 13:56:03 crc kubenswrapper[4815]: I0225 13:56:03.296203 4815 generic.go:334] "Generic (PLEG): container finished" podID="7d5d295f-9999-4500-af75-c46724b26d89" containerID="7b9bbcacfb5d26bb4c9bad6ee1cd8df0a2010a679261982c3a3c7ddd27bdb0bd" exitCode=0 Feb 25 13:56:03 crc kubenswrapper[4815]: I0225 13:56:03.296279 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533796-68pz9" event={"ID":"7d5d295f-9999-4500-af75-c46724b26d89","Type":"ContainerDied","Data":"7b9bbcacfb5d26bb4c9bad6ee1cd8df0a2010a679261982c3a3c7ddd27bdb0bd"} Feb 25 13:56:04 crc kubenswrapper[4815]: I0225 13:56:04.763759 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533796-68pz9" Feb 25 13:56:04 crc kubenswrapper[4815]: I0225 13:56:04.863995 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x82kl\" (UniqueName: \"kubernetes.io/projected/7d5d295f-9999-4500-af75-c46724b26d89-kube-api-access-x82kl\") pod \"7d5d295f-9999-4500-af75-c46724b26d89\" (UID: \"7d5d295f-9999-4500-af75-c46724b26d89\") " Feb 25 13:56:04 crc kubenswrapper[4815]: I0225 13:56:04.874893 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d5d295f-9999-4500-af75-c46724b26d89-kube-api-access-x82kl" (OuterVolumeSpecName: "kube-api-access-x82kl") pod "7d5d295f-9999-4500-af75-c46724b26d89" (UID: "7d5d295f-9999-4500-af75-c46724b26d89"). InnerVolumeSpecName "kube-api-access-x82kl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:56:04 crc kubenswrapper[4815]: I0225 13:56:04.967952 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x82kl\" (UniqueName: \"kubernetes.io/projected/7d5d295f-9999-4500-af75-c46724b26d89-kube-api-access-x82kl\") on node \"crc\" DevicePath \"\"" Feb 25 13:56:05 crc kubenswrapper[4815]: I0225 13:56:05.321011 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533796-68pz9" event={"ID":"7d5d295f-9999-4500-af75-c46724b26d89","Type":"ContainerDied","Data":"2d3e3870cc305292e4febac36f113f94c247b6b511226e13b95bbd5cc90a531b"} Feb 25 13:56:05 crc kubenswrapper[4815]: I0225 13:56:05.321604 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d3e3870cc305292e4febac36f113f94c247b6b511226e13b95bbd5cc90a531b" Feb 25 13:56:05 crc kubenswrapper[4815]: I0225 13:56:05.321130 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533796-68pz9" Feb 25 13:56:05 crc kubenswrapper[4815]: I0225 13:56:05.861977 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533790-6cjzj"] Feb 25 13:56:05 crc kubenswrapper[4815]: I0225 13:56:05.874588 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533790-6cjzj"] Feb 25 13:56:06 crc kubenswrapper[4815]: I0225 13:56:06.949001 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="893e8504-9fb6-4734-aa82-aac7dfecb7c7" path="/var/lib/kubelet/pods/893e8504-9fb6-4734-aa82-aac7dfecb7c7/volumes" Feb 25 13:56:06 crc kubenswrapper[4815]: I0225 13:56:06.989072 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tbhrj"] Feb 25 13:56:06 crc kubenswrapper[4815]: E0225 13:56:06.989639 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5d295f-9999-4500-af75-c46724b26d89" containerName="oc" Feb 25 13:56:06 crc kubenswrapper[4815]: I0225 13:56:06.989658 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5d295f-9999-4500-af75-c46724b26d89" containerName="oc" Feb 25 13:56:06 crc kubenswrapper[4815]: I0225 13:56:06.989888 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5d295f-9999-4500-af75-c46724b26d89" containerName="oc" Feb 25 13:56:06 crc kubenswrapper[4815]: I0225 13:56:06.991382 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tbhrj" Feb 25 13:56:07 crc kubenswrapper[4815]: I0225 13:56:07.000768 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tbhrj"] Feb 25 13:56:07 crc kubenswrapper[4815]: I0225 13:56:07.017407 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-catalog-content\") pod \"redhat-operators-tbhrj\" (UID: \"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881\") " pod="openshift-marketplace/redhat-operators-tbhrj" Feb 25 13:56:07 crc kubenswrapper[4815]: I0225 13:56:07.017460 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdzl9\" (UniqueName: \"kubernetes.io/projected/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-kube-api-access-fdzl9\") pod \"redhat-operators-tbhrj\" (UID: \"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881\") " pod="openshift-marketplace/redhat-operators-tbhrj" Feb 25 13:56:07 crc kubenswrapper[4815]: I0225 13:56:07.017594 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-utilities\") pod \"redhat-operators-tbhrj\" (UID: \"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881\") " pod="openshift-marketplace/redhat-operators-tbhrj" Feb 25 13:56:07 crc kubenswrapper[4815]: I0225 13:56:07.165917 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-catalog-content\") pod \"redhat-operators-tbhrj\" (UID: \"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881\") " pod="openshift-marketplace/redhat-operators-tbhrj" Feb 25 13:56:07 crc kubenswrapper[4815]: I0225 13:56:07.165978 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdzl9\" (UniqueName: \"kubernetes.io/projected/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-kube-api-access-fdzl9\") pod \"redhat-operators-tbhrj\" (UID: \"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881\") " pod="openshift-marketplace/redhat-operators-tbhrj" Feb 25 13:56:07 crc kubenswrapper[4815]: I0225 13:56:07.166093 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-utilities\") pod \"redhat-operators-tbhrj\" (UID: \"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881\") " pod="openshift-marketplace/redhat-operators-tbhrj" Feb 25 13:56:07 crc kubenswrapper[4815]: I0225 13:56:07.167080 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-catalog-content\") pod \"redhat-operators-tbhrj\" (UID: \"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881\") " pod="openshift-marketplace/redhat-operators-tbhrj" Feb 25 13:56:07 crc kubenswrapper[4815]: I0225 13:56:07.167168 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-utilities\") pod \"redhat-operators-tbhrj\" (UID: \"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881\") " pod="openshift-marketplace/redhat-operators-tbhrj" Feb 25 13:56:07 crc kubenswrapper[4815]: I0225 13:56:07.197458 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdzl9\" (UniqueName: \"kubernetes.io/projected/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-kube-api-access-fdzl9\") pod \"redhat-operators-tbhrj\" (UID: \"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881\") " pod="openshift-marketplace/redhat-operators-tbhrj" Feb 25 13:56:07 crc kubenswrapper[4815]: I0225 13:56:07.318760 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tbhrj" Feb 25 13:56:07 crc kubenswrapper[4815]: I0225 13:56:07.881534 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tbhrj"] Feb 25 13:56:07 crc kubenswrapper[4815]: W0225 13:56:07.886269 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59a5ab31_9dce_4fc3_b0ae_fe251e9c3881.slice/crio-0ec52def9053c6bc97fc2879d95fe8148393bc24700116ae19ae8fe6e015f48a WatchSource:0}: Error finding container 0ec52def9053c6bc97fc2879d95fe8148393bc24700116ae19ae8fe6e015f48a: Status 404 returned error can't find the container with id 0ec52def9053c6bc97fc2879d95fe8148393bc24700116ae19ae8fe6e015f48a Feb 25 13:56:08 crc kubenswrapper[4815]: I0225 13:56:08.350851 4815 generic.go:334] "Generic (PLEG): container finished" podID="59a5ab31-9dce-4fc3-b0ae-fe251e9c3881" containerID="1da419c2f97c8ef19d7b862361945da6c90cc7f2f469387579b550db13c27a94" exitCode=0 Feb 25 13:56:08 crc kubenswrapper[4815]: I0225 13:56:08.350903 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbhrj" event={"ID":"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881","Type":"ContainerDied","Data":"1da419c2f97c8ef19d7b862361945da6c90cc7f2f469387579b550db13c27a94"} Feb 25 13:56:08 crc kubenswrapper[4815]: I0225 13:56:08.350937 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbhrj" event={"ID":"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881","Type":"ContainerStarted","Data":"0ec52def9053c6bc97fc2879d95fe8148393bc24700116ae19ae8fe6e015f48a"} Feb 25 13:56:10 crc kubenswrapper[4815]: I0225 13:56:10.373621 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbhrj" event={"ID":"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881","Type":"ContainerStarted","Data":"0eb5bda6bd9c416a0338cc53f67c48f8c8fb2322c30a78c4482e09ddfb75fcc8"} Feb 25 13:56:12 crc kubenswrapper[4815]: I0225 13:56:12.394767 4815 generic.go:334] "Generic (PLEG): container finished" podID="59a5ab31-9dce-4fc3-b0ae-fe251e9c3881" containerID="0eb5bda6bd9c416a0338cc53f67c48f8c8fb2322c30a78c4482e09ddfb75fcc8" exitCode=0 Feb 25 13:56:12 crc kubenswrapper[4815]: I0225 13:56:12.394840 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbhrj" event={"ID":"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881","Type":"ContainerDied","Data":"0eb5bda6bd9c416a0338cc53f67c48f8c8fb2322c30a78c4482e09ddfb75fcc8"} Feb 25 13:56:13 crc kubenswrapper[4815]: I0225 13:56:13.412127 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbhrj" event={"ID":"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881","Type":"ContainerStarted","Data":"347e5033a5eb2457efa107d2602239d4ff774aabb40651a0ac1b916d8a6f509b"} Feb 25 13:56:13 crc kubenswrapper[4815]: I0225 13:56:13.432753 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tbhrj" podStartSLOduration=2.868260898 podStartE2EDuration="7.43272902s" podCreationTimestamp="2026-02-25 13:56:06 +0000 UTC" firstStartedPulling="2026-02-25 13:56:08.352976 +0000 UTC m=+2126.154074054" lastFinishedPulling="2026-02-25 13:56:12.917444112 +0000 UTC m=+2130.718542176" observedRunningTime="2026-02-25 13:56:13.429991524 +0000 UTC m=+2131.231089578" watchObservedRunningTime="2026-02-25 13:56:13.43272902 +0000 UTC m=+2131.233827094" Feb 25 13:56:17 crc kubenswrapper[4815]: I0225 13:56:17.319936 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tbhrj" Feb 25 13:56:17 crc kubenswrapper[4815]: I0225 13:56:17.320747 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tbhrj" Feb 25 13:56:18 crc kubenswrapper[4815]: I0225 13:56:18.410724 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tbhrj" podUID="59a5ab31-9dce-4fc3-b0ae-fe251e9c3881" containerName="registry-server" probeResult="failure" output=< Feb 25 13:56:18 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Feb 25 13:56:18 crc kubenswrapper[4815]: > Feb 25 13:56:27 crc kubenswrapper[4815]: I0225 13:56:27.395188 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tbhrj" Feb 25 13:56:27 crc kubenswrapper[4815]: I0225 13:56:27.475584 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tbhrj" Feb 25 13:56:27 crc kubenswrapper[4815]: I0225 13:56:27.643189 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tbhrj"] Feb 25 13:56:28 crc kubenswrapper[4815]: I0225 13:56:28.559748 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tbhrj" podUID="59a5ab31-9dce-4fc3-b0ae-fe251e9c3881" containerName="registry-server" containerID="cri-o://347e5033a5eb2457efa107d2602239d4ff774aabb40651a0ac1b916d8a6f509b" gracePeriod=2 Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.133294 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tbhrj" Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.182653 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdzl9\" (UniqueName: \"kubernetes.io/projected/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-kube-api-access-fdzl9\") pod \"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881\" (UID: \"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881\") " Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.182756 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-utilities\") pod \"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881\" (UID: \"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881\") " Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.183071 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-catalog-content\") pod \"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881\" (UID: \"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881\") " Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.188778 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-utilities" (OuterVolumeSpecName: "utilities") pod "59a5ab31-9dce-4fc3-b0ae-fe251e9c3881" (UID: "59a5ab31-9dce-4fc3-b0ae-fe251e9c3881"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.194155 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-kube-api-access-fdzl9" (OuterVolumeSpecName: "kube-api-access-fdzl9") pod "59a5ab31-9dce-4fc3-b0ae-fe251e9c3881" (UID: "59a5ab31-9dce-4fc3-b0ae-fe251e9c3881"). InnerVolumeSpecName "kube-api-access-fdzl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.285713 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdzl9\" (UniqueName: \"kubernetes.io/projected/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-kube-api-access-fdzl9\") on node \"crc\" DevicePath \"\"" Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.286672 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.336834 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59a5ab31-9dce-4fc3-b0ae-fe251e9c3881" (UID: "59a5ab31-9dce-4fc3-b0ae-fe251e9c3881"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.388357 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.572932 4815 generic.go:334] "Generic (PLEG): container finished" podID="59a5ab31-9dce-4fc3-b0ae-fe251e9c3881" containerID="347e5033a5eb2457efa107d2602239d4ff774aabb40651a0ac1b916d8a6f509b" exitCode=0 Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.572979 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbhrj" event={"ID":"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881","Type":"ContainerDied","Data":"347e5033a5eb2457efa107d2602239d4ff774aabb40651a0ac1b916d8a6f509b"} Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.572990 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tbhrj" Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.573009 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbhrj" event={"ID":"59a5ab31-9dce-4fc3-b0ae-fe251e9c3881","Type":"ContainerDied","Data":"0ec52def9053c6bc97fc2879d95fe8148393bc24700116ae19ae8fe6e015f48a"} Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.573029 4815 scope.go:117] "RemoveContainer" containerID="347e5033a5eb2457efa107d2602239d4ff774aabb40651a0ac1b916d8a6f509b" Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.611995 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tbhrj"] Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.613298 4815 scope.go:117] "RemoveContainer" containerID="0eb5bda6bd9c416a0338cc53f67c48f8c8fb2322c30a78c4482e09ddfb75fcc8" Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.620173 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tbhrj"] Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.647798 4815 scope.go:117] "RemoveContainer" containerID="1da419c2f97c8ef19d7b862361945da6c90cc7f2f469387579b550db13c27a94" Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.675756 4815 scope.go:117] "RemoveContainer" containerID="347e5033a5eb2457efa107d2602239d4ff774aabb40651a0ac1b916d8a6f509b" Feb 25 13:56:29 crc kubenswrapper[4815]: E0225 13:56:29.676254 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"347e5033a5eb2457efa107d2602239d4ff774aabb40651a0ac1b916d8a6f509b\": container with ID starting with 347e5033a5eb2457efa107d2602239d4ff774aabb40651a0ac1b916d8a6f509b not found: ID does not exist" containerID="347e5033a5eb2457efa107d2602239d4ff774aabb40651a0ac1b916d8a6f509b" Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.676289 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"347e5033a5eb2457efa107d2602239d4ff774aabb40651a0ac1b916d8a6f509b"} err="failed to get container status \"347e5033a5eb2457efa107d2602239d4ff774aabb40651a0ac1b916d8a6f509b\": rpc error: code = NotFound desc = could not find container \"347e5033a5eb2457efa107d2602239d4ff774aabb40651a0ac1b916d8a6f509b\": container with ID starting with 347e5033a5eb2457efa107d2602239d4ff774aabb40651a0ac1b916d8a6f509b not found: ID does not exist" Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.676308 4815 scope.go:117] "RemoveContainer" containerID="0eb5bda6bd9c416a0338cc53f67c48f8c8fb2322c30a78c4482e09ddfb75fcc8" Feb 25 13:56:29 crc kubenswrapper[4815]: E0225 13:56:29.677345 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eb5bda6bd9c416a0338cc53f67c48f8c8fb2322c30a78c4482e09ddfb75fcc8\": container with ID starting with 0eb5bda6bd9c416a0338cc53f67c48f8c8fb2322c30a78c4482e09ddfb75fcc8 not found: ID does not exist" containerID="0eb5bda6bd9c416a0338cc53f67c48f8c8fb2322c30a78c4482e09ddfb75fcc8" Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.677400 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eb5bda6bd9c416a0338cc53f67c48f8c8fb2322c30a78c4482e09ddfb75fcc8"} err="failed to get container status \"0eb5bda6bd9c416a0338cc53f67c48f8c8fb2322c30a78c4482e09ddfb75fcc8\": rpc error: code = NotFound desc = could not find container \"0eb5bda6bd9c416a0338cc53f67c48f8c8fb2322c30a78c4482e09ddfb75fcc8\": container with ID starting with 0eb5bda6bd9c416a0338cc53f67c48f8c8fb2322c30a78c4482e09ddfb75fcc8 not found: ID does not exist" Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.677437 4815 scope.go:117] "RemoveContainer" containerID="1da419c2f97c8ef19d7b862361945da6c90cc7f2f469387579b550db13c27a94" Feb 25 13:56:29 crc kubenswrapper[4815]: E0225 13:56:29.677868 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1da419c2f97c8ef19d7b862361945da6c90cc7f2f469387579b550db13c27a94\": container with ID starting with 1da419c2f97c8ef19d7b862361945da6c90cc7f2f469387579b550db13c27a94 not found: ID does not exist" containerID="1da419c2f97c8ef19d7b862361945da6c90cc7f2f469387579b550db13c27a94" Feb 25 13:56:29 crc kubenswrapper[4815]: I0225 13:56:29.677895 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1da419c2f97c8ef19d7b862361945da6c90cc7f2f469387579b550db13c27a94"} err="failed to get container status \"1da419c2f97c8ef19d7b862361945da6c90cc7f2f469387579b550db13c27a94\": rpc error: code = NotFound desc = could not find container \"1da419c2f97c8ef19d7b862361945da6c90cc7f2f469387579b550db13c27a94\": container with ID starting with 1da419c2f97c8ef19d7b862361945da6c90cc7f2f469387579b550db13c27a94 not found: ID does not exist" Feb 25 13:56:30 crc kubenswrapper[4815]: I0225 13:56:30.956155 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59a5ab31-9dce-4fc3-b0ae-fe251e9c3881" path="/var/lib/kubelet/pods/59a5ab31-9dce-4fc3-b0ae-fe251e9c3881/volumes" Feb 25 13:56:48 crc kubenswrapper[4815]: I0225 13:56:48.417690 4815 scope.go:117] "RemoveContainer" containerID="88339ec1ab335acc17c088dc3f0fd6d9ce7ab5947f12755f719e8a51056f92c1" Feb 25 13:56:51 crc kubenswrapper[4815]: I0225 13:56:51.342436 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:56:51 crc kubenswrapper[4815]: I0225 13:56:51.343042 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:57:08 crc kubenswrapper[4815]: I0225 13:57:08.981892 4815 generic.go:334] "Generic (PLEG): container finished" podID="4e6edebf-da64-4665-9781-b4f3191e0d10" containerID="ffbed81d33216662483df14d093ab90e9bfc9726d21a12a22093bc0fdaffbb08" exitCode=0 Feb 25 13:57:08 crc kubenswrapper[4815]: I0225 13:57:08.982547 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" event={"ID":"4e6edebf-da64-4665-9781-b4f3191e0d10","Type":"ContainerDied","Data":"ffbed81d33216662483df14d093ab90e9bfc9726d21a12a22093bc0fdaffbb08"} Feb 25 13:57:10 crc kubenswrapper[4815]: I0225 13:57:10.396171 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:57:10 crc kubenswrapper[4815]: I0225 13:57:10.548917 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-inventory\") pod \"4e6edebf-da64-4665-9781-b4f3191e0d10\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " Feb 25 13:57:10 crc kubenswrapper[4815]: I0225 13:57:10.549233 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-libvirt-secret-0\") pod \"4e6edebf-da64-4665-9781-b4f3191e0d10\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " Feb 25 13:57:10 crc kubenswrapper[4815]: I0225 13:57:10.549916 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrwgb\" (UniqueName: \"kubernetes.io/projected/4e6edebf-da64-4665-9781-b4f3191e0d10-kube-api-access-mrwgb\") pod \"4e6edebf-da64-4665-9781-b4f3191e0d10\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " Feb 25 13:57:10 crc kubenswrapper[4815]: I0225 13:57:10.550005 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-libvirt-combined-ca-bundle\") pod \"4e6edebf-da64-4665-9781-b4f3191e0d10\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " Feb 25 13:57:10 crc kubenswrapper[4815]: I0225 13:57:10.550031 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-ssh-key-openstack-edpm-ipam\") pod \"4e6edebf-da64-4665-9781-b4f3191e0d10\" (UID: \"4e6edebf-da64-4665-9781-b4f3191e0d10\") " Feb 25 13:57:10 crc kubenswrapper[4815]: I0225 13:57:10.556099 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e6edebf-da64-4665-9781-b4f3191e0d10-kube-api-access-mrwgb" (OuterVolumeSpecName: "kube-api-access-mrwgb") pod "4e6edebf-da64-4665-9781-b4f3191e0d10" (UID: "4e6edebf-da64-4665-9781-b4f3191e0d10"). InnerVolumeSpecName "kube-api-access-mrwgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:57:10 crc kubenswrapper[4815]: I0225 13:57:10.558742 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "4e6edebf-da64-4665-9781-b4f3191e0d10" (UID: "4e6edebf-da64-4665-9781-b4f3191e0d10"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:57:10 crc kubenswrapper[4815]: I0225 13:57:10.577601 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "4e6edebf-da64-4665-9781-b4f3191e0d10" (UID: "4e6edebf-da64-4665-9781-b4f3191e0d10"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:57:10 crc kubenswrapper[4815]: I0225 13:57:10.582879 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4e6edebf-da64-4665-9781-b4f3191e0d10" (UID: "4e6edebf-da64-4665-9781-b4f3191e0d10"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:57:10 crc kubenswrapper[4815]: I0225 13:57:10.584318 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-inventory" (OuterVolumeSpecName: "inventory") pod "4e6edebf-da64-4665-9781-b4f3191e0d10" (UID: "4e6edebf-da64-4665-9781-b4f3191e0d10"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:57:10 crc kubenswrapper[4815]: I0225 13:57:10.652317 4815 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:57:10 crc kubenswrapper[4815]: I0225 13:57:10.652362 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrwgb\" (UniqueName: \"kubernetes.io/projected/4e6edebf-da64-4665-9781-b4f3191e0d10-kube-api-access-mrwgb\") on node \"crc\" DevicePath \"\"" Feb 25 13:57:10 crc kubenswrapper[4815]: I0225 13:57:10.652379 4815 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:57:10 crc kubenswrapper[4815]: I0225 13:57:10.652394 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 13:57:10 crc kubenswrapper[4815]: I0225 13:57:10.652407 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e6edebf-da64-4665-9781-b4f3191e0d10-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.011796 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" event={"ID":"4e6edebf-da64-4665-9781-b4f3191e0d10","Type":"ContainerDied","Data":"07ca3b66ec0c4ae425a50946b5e980c3f4804b97e2092340685bcbc9468dd4b0"} Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.012076 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-85zsm" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.013603 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07ca3b66ec0c4ae425a50946b5e980c3f4804b97e2092340685bcbc9468dd4b0" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.106724 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6"] Feb 25 13:57:11 crc kubenswrapper[4815]: E0225 13:57:11.107184 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59a5ab31-9dce-4fc3-b0ae-fe251e9c3881" containerName="extract-utilities" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.107205 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="59a5ab31-9dce-4fc3-b0ae-fe251e9c3881" containerName="extract-utilities" Feb 25 13:57:11 crc kubenswrapper[4815]: E0225 13:57:11.107230 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59a5ab31-9dce-4fc3-b0ae-fe251e9c3881" containerName="extract-content" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.107239 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="59a5ab31-9dce-4fc3-b0ae-fe251e9c3881" containerName="extract-content" Feb 25 13:57:11 crc kubenswrapper[4815]: E0225 13:57:11.107273 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59a5ab31-9dce-4fc3-b0ae-fe251e9c3881" containerName="registry-server" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.107282 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="59a5ab31-9dce-4fc3-b0ae-fe251e9c3881" containerName="registry-server" Feb 25 13:57:11 crc kubenswrapper[4815]: E0225 13:57:11.107296 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e6edebf-da64-4665-9781-b4f3191e0d10" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.107304 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e6edebf-da64-4665-9781-b4f3191e0d10" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.107539 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e6edebf-da64-4665-9781-b4f3191e0d10" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.107569 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="59a5ab31-9dce-4fc3-b0ae-fe251e9c3881" containerName="registry-server" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.108263 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.110782 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.111060 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.111180 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.111285 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mmbj" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.111767 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.117080 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6"] Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.139346 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.139658 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.264139 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.264246 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.264272 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d96r7\" (UniqueName: \"kubernetes.io/projected/84065eef-8fae-4e48-87fa-900a2eab8748-kube-api-access-d96r7\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.264296 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.264332 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.264407 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.264431 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.264467 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/84065eef-8fae-4e48-87fa-900a2eab8748-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.264586 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.264609 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.264714 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.366416 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/84065eef-8fae-4e48-87fa-900a2eab8748-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.366521 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.366559 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.366638 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.366700 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.366736 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.366759 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d96r7\" (UniqueName: \"kubernetes.io/projected/84065eef-8fae-4e48-87fa-900a2eab8748-kube-api-access-d96r7\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.366781 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.366810 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.366852 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.366873 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.367815 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/84065eef-8fae-4e48-87fa-900a2eab8748-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.371794 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.372059 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.372607 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.372731 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.372830 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.379059 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.379167 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.379247 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.379293 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.389572 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d96r7\" (UniqueName: \"kubernetes.io/projected/84065eef-8fae-4e48-87fa-900a2eab8748-kube-api-access-d96r7\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kttx6\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:11 crc kubenswrapper[4815]: I0225 13:57:11.471007 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:57:12 crc kubenswrapper[4815]: I0225 13:57:12.002651 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6"] Feb 25 13:57:12 crc kubenswrapper[4815]: I0225 13:57:12.022055 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" event={"ID":"84065eef-8fae-4e48-87fa-900a2eab8748","Type":"ContainerStarted","Data":"c9a70e37a6aed67f3c4da8e37a262f69171f53b7e3a08ad46200cec483c23c3a"} Feb 25 13:57:13 crc kubenswrapper[4815]: I0225 13:57:13.031446 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" event={"ID":"84065eef-8fae-4e48-87fa-900a2eab8748","Type":"ContainerStarted","Data":"642d146a80e368ac9338cc3a1e4e21408e6e3705b759c06cca04a5e80979efb5"} Feb 25 13:57:21 crc kubenswrapper[4815]: I0225 13:57:21.342329 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:57:21 crc kubenswrapper[4815]: I0225 13:57:21.343245 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:57:51 crc kubenswrapper[4815]: I0225 13:57:51.341966 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 13:57:51 crc kubenswrapper[4815]: I0225 13:57:51.342858 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 13:57:51 crc kubenswrapper[4815]: I0225 13:57:51.342913 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 13:57:51 crc kubenswrapper[4815]: I0225 13:57:51.343814 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05"} pod="openshift-machine-config-operator/machine-config-daemon-frc75" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 13:57:51 crc kubenswrapper[4815]: I0225 13:57:51.343914 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" containerID="cri-o://9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" gracePeriod=600 Feb 25 13:57:51 crc kubenswrapper[4815]: E0225 13:57:51.464139 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:57:52 crc kubenswrapper[4815]: I0225 13:57:52.442062 4815 generic.go:334] "Generic (PLEG): container finished" podID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" exitCode=0 Feb 25 13:57:52 crc kubenswrapper[4815]: I0225 13:57:52.442375 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerDied","Data":"9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05"} Feb 25 13:57:52 crc kubenswrapper[4815]: I0225 13:57:52.442417 4815 scope.go:117] "RemoveContainer" containerID="bc1a87ba95ff2fe57f3edfcdf517b6e2b034a474d761b615e638c4a52c55574e" Feb 25 13:57:52 crc kubenswrapper[4815]: I0225 13:57:52.443123 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 13:57:52 crc kubenswrapper[4815]: E0225 13:57:52.443408 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:57:52 crc kubenswrapper[4815]: I0225 13:57:52.462594 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" podStartSLOduration=40.917393191 podStartE2EDuration="41.462576697s" podCreationTimestamp="2026-02-25 13:57:11 +0000 UTC" firstStartedPulling="2026-02-25 13:57:12.013636386 +0000 UTC m=+2189.814734440" lastFinishedPulling="2026-02-25 13:57:12.558819892 +0000 UTC m=+2190.359917946" observedRunningTime="2026-02-25 13:57:13.061360401 +0000 UTC m=+2190.862458465" watchObservedRunningTime="2026-02-25 13:57:52.462576697 +0000 UTC m=+2230.263674751" Feb 25 13:58:00 crc kubenswrapper[4815]: I0225 13:58:00.149705 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533798-7bnr8"] Feb 25 13:58:00 crc kubenswrapper[4815]: I0225 13:58:00.151756 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533798-7bnr8" Feb 25 13:58:00 crc kubenswrapper[4815]: I0225 13:58:00.160097 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 13:58:00 crc kubenswrapper[4815]: I0225 13:58:00.160655 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 13:58:00 crc kubenswrapper[4815]: I0225 13:58:00.160982 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 13:58:00 crc kubenswrapper[4815]: I0225 13:58:00.162813 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533798-7bnr8"] Feb 25 13:58:00 crc kubenswrapper[4815]: I0225 13:58:00.270438 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqcpc\" (UniqueName: \"kubernetes.io/projected/7e7253d5-5810-4271-a58b-040a3eb09d82-kube-api-access-jqcpc\") pod \"auto-csr-approver-29533798-7bnr8\" (UID: \"7e7253d5-5810-4271-a58b-040a3eb09d82\") " pod="openshift-infra/auto-csr-approver-29533798-7bnr8" Feb 25 13:58:00 crc kubenswrapper[4815]: I0225 13:58:00.371966 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqcpc\" (UniqueName: \"kubernetes.io/projected/7e7253d5-5810-4271-a58b-040a3eb09d82-kube-api-access-jqcpc\") pod \"auto-csr-approver-29533798-7bnr8\" (UID: \"7e7253d5-5810-4271-a58b-040a3eb09d82\") " pod="openshift-infra/auto-csr-approver-29533798-7bnr8" Feb 25 13:58:00 crc kubenswrapper[4815]: I0225 13:58:00.400713 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqcpc\" (UniqueName: \"kubernetes.io/projected/7e7253d5-5810-4271-a58b-040a3eb09d82-kube-api-access-jqcpc\") pod \"auto-csr-approver-29533798-7bnr8\" (UID: \"7e7253d5-5810-4271-a58b-040a3eb09d82\") " pod="openshift-infra/auto-csr-approver-29533798-7bnr8" Feb 25 13:58:00 crc kubenswrapper[4815]: I0225 13:58:00.476640 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533798-7bnr8" Feb 25 13:58:00 crc kubenswrapper[4815]: I0225 13:58:00.950552 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533798-7bnr8"] Feb 25 13:58:01 crc kubenswrapper[4815]: I0225 13:58:01.529312 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533798-7bnr8" event={"ID":"7e7253d5-5810-4271-a58b-040a3eb09d82","Type":"ContainerStarted","Data":"b6e9c50c88a52842007aa9fc2236d878d2cfdcda4c586e2729ff6237ccb19cb8"} Feb 25 13:58:02 crc kubenswrapper[4815]: I0225 13:58:02.541920 4815 generic.go:334] "Generic (PLEG): container finished" podID="7e7253d5-5810-4271-a58b-040a3eb09d82" containerID="c94113bf1e90f8bff4a9480c92916ba73c04497222227dba629dc1ea749df28b" exitCode=0 Feb 25 13:58:02 crc kubenswrapper[4815]: I0225 13:58:02.542015 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533798-7bnr8" event={"ID":"7e7253d5-5810-4271-a58b-040a3eb09d82","Type":"ContainerDied","Data":"c94113bf1e90f8bff4a9480c92916ba73c04497222227dba629dc1ea749df28b"} Feb 25 13:58:03 crc kubenswrapper[4815]: I0225 13:58:03.833941 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533798-7bnr8" Feb 25 13:58:03 crc kubenswrapper[4815]: I0225 13:58:03.940643 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqcpc\" (UniqueName: \"kubernetes.io/projected/7e7253d5-5810-4271-a58b-040a3eb09d82-kube-api-access-jqcpc\") pod \"7e7253d5-5810-4271-a58b-040a3eb09d82\" (UID: \"7e7253d5-5810-4271-a58b-040a3eb09d82\") " Feb 25 13:58:03 crc kubenswrapper[4815]: I0225 13:58:03.946775 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e7253d5-5810-4271-a58b-040a3eb09d82-kube-api-access-jqcpc" (OuterVolumeSpecName: "kube-api-access-jqcpc") pod "7e7253d5-5810-4271-a58b-040a3eb09d82" (UID: "7e7253d5-5810-4271-a58b-040a3eb09d82"). InnerVolumeSpecName "kube-api-access-jqcpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:58:04 crc kubenswrapper[4815]: I0225 13:58:04.042439 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqcpc\" (UniqueName: \"kubernetes.io/projected/7e7253d5-5810-4271-a58b-040a3eb09d82-kube-api-access-jqcpc\") on node \"crc\" DevicePath \"\"" Feb 25 13:58:04 crc kubenswrapper[4815]: I0225 13:58:04.563796 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533798-7bnr8" event={"ID":"7e7253d5-5810-4271-a58b-040a3eb09d82","Type":"ContainerDied","Data":"b6e9c50c88a52842007aa9fc2236d878d2cfdcda4c586e2729ff6237ccb19cb8"} Feb 25 13:58:04 crc kubenswrapper[4815]: I0225 13:58:04.564110 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6e9c50c88a52842007aa9fc2236d878d2cfdcda4c586e2729ff6237ccb19cb8" Feb 25 13:58:04 crc kubenswrapper[4815]: I0225 13:58:04.563881 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533798-7bnr8" Feb 25 13:58:04 crc kubenswrapper[4815]: I0225 13:58:04.923833 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533792-b2jvf"] Feb 25 13:58:04 crc kubenswrapper[4815]: I0225 13:58:04.935031 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533792-b2jvf"] Feb 25 13:58:04 crc kubenswrapper[4815]: I0225 13:58:04.948714 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4449c14e-2d79-4ce8-ba74-4368a26f5d44" path="/var/lib/kubelet/pods/4449c14e-2d79-4ce8-ba74-4368a26f5d44/volumes" Feb 25 13:58:06 crc kubenswrapper[4815]: I0225 13:58:06.935576 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 13:58:06 crc kubenswrapper[4815]: E0225 13:58:06.936064 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:58:17 crc kubenswrapper[4815]: I0225 13:58:17.936279 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 13:58:17 crc kubenswrapper[4815]: E0225 13:58:17.937109 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:58:31 crc kubenswrapper[4815]: I0225 13:58:31.936494 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 13:58:31 crc kubenswrapper[4815]: E0225 13:58:31.937550 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:58:44 crc kubenswrapper[4815]: I0225 13:58:44.937140 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 13:58:44 crc kubenswrapper[4815]: E0225 13:58:44.940167 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:58:48 crc kubenswrapper[4815]: I0225 13:58:48.521938 4815 scope.go:117] "RemoveContainer" containerID="d98d1003e94464d3ea2ae27c5ebb11913995cf8c0c5b029f9f8db5a79f4d35b2" Feb 25 13:58:58 crc kubenswrapper[4815]: I0225 13:58:58.935934 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 13:58:58 crc kubenswrapper[4815]: E0225 13:58:58.936903 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:59:11 crc kubenswrapper[4815]: I0225 13:59:11.936535 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 13:59:11 crc kubenswrapper[4815]: E0225 13:59:11.937660 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:59:24 crc kubenswrapper[4815]: I0225 13:59:24.936243 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 13:59:24 crc kubenswrapper[4815]: E0225 13:59:24.937079 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:59:29 crc kubenswrapper[4815]: I0225 13:59:29.430824 4815 generic.go:334] "Generic (PLEG): container finished" podID="84065eef-8fae-4e48-87fa-900a2eab8748" containerID="642d146a80e368ac9338cc3a1e4e21408e6e3705b759c06cca04a5e80979efb5" exitCode=0 Feb 25 13:59:29 crc kubenswrapper[4815]: I0225 13:59:29.430966 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" event={"ID":"84065eef-8fae-4e48-87fa-900a2eab8748","Type":"ContainerDied","Data":"642d146a80e368ac9338cc3a1e4e21408e6e3705b759c06cca04a5e80979efb5"} Feb 25 13:59:30 crc kubenswrapper[4815]: I0225 13:59:30.934032 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.002501 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-0\") pod \"84065eef-8fae-4e48-87fa-900a2eab8748\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.002598 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-ssh-key-openstack-edpm-ipam\") pod \"84065eef-8fae-4e48-87fa-900a2eab8748\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.002660 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-inventory\") pod \"84065eef-8fae-4e48-87fa-900a2eab8748\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.002702 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-migration-ssh-key-1\") pod \"84065eef-8fae-4e48-87fa-900a2eab8748\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.002732 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-2\") pod \"84065eef-8fae-4e48-87fa-900a2eab8748\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.002760 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-migration-ssh-key-0\") pod \"84065eef-8fae-4e48-87fa-900a2eab8748\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.002791 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d96r7\" (UniqueName: \"kubernetes.io/projected/84065eef-8fae-4e48-87fa-900a2eab8748-kube-api-access-d96r7\") pod \"84065eef-8fae-4e48-87fa-900a2eab8748\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.002867 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-3\") pod \"84065eef-8fae-4e48-87fa-900a2eab8748\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.002905 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/84065eef-8fae-4e48-87fa-900a2eab8748-nova-extra-config-0\") pod \"84065eef-8fae-4e48-87fa-900a2eab8748\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.002924 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-1\") pod \"84065eef-8fae-4e48-87fa-900a2eab8748\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.002948 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-combined-ca-bundle\") pod \"84065eef-8fae-4e48-87fa-900a2eab8748\" (UID: \"84065eef-8fae-4e48-87fa-900a2eab8748\") " Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.008672 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84065eef-8fae-4e48-87fa-900a2eab8748-kube-api-access-d96r7" (OuterVolumeSpecName: "kube-api-access-d96r7") pod "84065eef-8fae-4e48-87fa-900a2eab8748" (UID: "84065eef-8fae-4e48-87fa-900a2eab8748"). InnerVolumeSpecName "kube-api-access-d96r7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.008745 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "84065eef-8fae-4e48-87fa-900a2eab8748" (UID: "84065eef-8fae-4e48-87fa-900a2eab8748"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.030747 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "84065eef-8fae-4e48-87fa-900a2eab8748" (UID: "84065eef-8fae-4e48-87fa-900a2eab8748"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.033159 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "84065eef-8fae-4e48-87fa-900a2eab8748" (UID: "84065eef-8fae-4e48-87fa-900a2eab8748"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.039437 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "84065eef-8fae-4e48-87fa-900a2eab8748" (UID: "84065eef-8fae-4e48-87fa-900a2eab8748"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.039636 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "84065eef-8fae-4e48-87fa-900a2eab8748" (UID: "84065eef-8fae-4e48-87fa-900a2eab8748"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.041344 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "84065eef-8fae-4e48-87fa-900a2eab8748" (UID: "84065eef-8fae-4e48-87fa-900a2eab8748"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.043070 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "84065eef-8fae-4e48-87fa-900a2eab8748" (UID: "84065eef-8fae-4e48-87fa-900a2eab8748"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.045876 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-inventory" (OuterVolumeSpecName: "inventory") pod "84065eef-8fae-4e48-87fa-900a2eab8748" (UID: "84065eef-8fae-4e48-87fa-900a2eab8748"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.051163 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84065eef-8fae-4e48-87fa-900a2eab8748-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "84065eef-8fae-4e48-87fa-900a2eab8748" (UID: "84065eef-8fae-4e48-87fa-900a2eab8748"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.057074 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "84065eef-8fae-4e48-87fa-900a2eab8748" (UID: "84065eef-8fae-4e48-87fa-900a2eab8748"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.106126 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.106171 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.106185 4815 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.106201 4815 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.106214 4815 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.106227 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d96r7\" (UniqueName: \"kubernetes.io/projected/84065eef-8fae-4e48-87fa-900a2eab8748-kube-api-access-d96r7\") on node \"crc\" DevicePath \"\"" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.106239 4815 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.106251 4815 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/84065eef-8fae-4e48-87fa-900a2eab8748-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.106263 4815 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.106278 4815 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.106289 4815 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/84065eef-8fae-4e48-87fa-900a2eab8748-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.451866 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" event={"ID":"84065eef-8fae-4e48-87fa-900a2eab8748","Type":"ContainerDied","Data":"c9a70e37a6aed67f3c4da8e37a262f69171f53b7e3a08ad46200cec483c23c3a"} Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.451909 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9a70e37a6aed67f3c4da8e37a262f69171f53b7e3a08ad46200cec483c23c3a" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.451970 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kttx6" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.586779 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh"] Feb 25 13:59:31 crc kubenswrapper[4815]: E0225 13:59:31.587208 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84065eef-8fae-4e48-87fa-900a2eab8748" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.587229 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="84065eef-8fae-4e48-87fa-900a2eab8748" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 25 13:59:31 crc kubenswrapper[4815]: E0225 13:59:31.587249 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e7253d5-5810-4271-a58b-040a3eb09d82" containerName="oc" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.587258 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e7253d5-5810-4271-a58b-040a3eb09d82" containerName="oc" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.587469 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e7253d5-5810-4271-a58b-040a3eb09d82" containerName="oc" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.587531 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="84065eef-8fae-4e48-87fa-900a2eab8748" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.588253 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.590702 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.591016 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.591665 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4mmbj" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.592383 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.593483 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.633908 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh"] Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.659205 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.659256 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.659442 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkfdw\" (UniqueName: \"kubernetes.io/projected/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-kube-api-access-dkfdw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.659524 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.659624 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.659679 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.659750 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.761333 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.761382 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.761462 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkfdw\" (UniqueName: \"kubernetes.io/projected/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-kube-api-access-dkfdw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.761526 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.761578 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.761615 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.761653 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.767276 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.768143 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.768759 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.768786 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.768783 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.769483 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.780895 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkfdw\" (UniqueName: \"kubernetes.io/projected/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-kube-api-access-dkfdw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-47nxh\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:31 crc kubenswrapper[4815]: I0225 13:59:31.907804 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 13:59:32 crc kubenswrapper[4815]: I0225 13:59:32.440950 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh"] Feb 25 13:59:32 crc kubenswrapper[4815]: W0225 13:59:32.441744 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b1ea76e_0503_4e5f_8adb_ad5e39e5bea6.slice/crio-d3a1f5d845196ee333027d9c993ad2211817eb1109c7ec059a527f87fb209d20 WatchSource:0}: Error finding container d3a1f5d845196ee333027d9c993ad2211817eb1109c7ec059a527f87fb209d20: Status 404 returned error can't find the container with id d3a1f5d845196ee333027d9c993ad2211817eb1109c7ec059a527f87fb209d20 Feb 25 13:59:32 crc kubenswrapper[4815]: I0225 13:59:32.463563 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" event={"ID":"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6","Type":"ContainerStarted","Data":"d3a1f5d845196ee333027d9c993ad2211817eb1109c7ec059a527f87fb209d20"} Feb 25 13:59:33 crc kubenswrapper[4815]: I0225 13:59:33.477666 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" event={"ID":"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6","Type":"ContainerStarted","Data":"afc4247d75b19ec77a9d3cb63c1fc28ca373ab90e03b0266b73d98ab4d113a48"} Feb 25 13:59:33 crc kubenswrapper[4815]: I0225 13:59:33.504145 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" podStartSLOduration=2.013420925 podStartE2EDuration="2.504123397s" podCreationTimestamp="2026-02-25 13:59:31 +0000 UTC" firstStartedPulling="2026-02-25 13:59:32.445143121 +0000 UTC m=+2330.246241175" lastFinishedPulling="2026-02-25 13:59:32.935845583 +0000 UTC m=+2330.736943647" observedRunningTime="2026-02-25 13:59:33.503212269 +0000 UTC m=+2331.304310363" watchObservedRunningTime="2026-02-25 13:59:33.504123397 +0000 UTC m=+2331.305221461" Feb 25 13:59:35 crc kubenswrapper[4815]: I0225 13:59:35.936459 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 13:59:35 crc kubenswrapper[4815]: E0225 13:59:35.937331 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 13:59:49 crc kubenswrapper[4815]: I0225 13:59:49.935332 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 13:59:49 crc kubenswrapper[4815]: E0225 13:59:49.936258 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.160285 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq"] Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.162632 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.171051 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.171219 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.172688 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533800-6drfm"] Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.174658 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533800-6drfm" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.176770 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.177398 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.177654 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.182931 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq"] Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.203338 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533800-6drfm"] Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.310430 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-config-volume\") pod \"collect-profiles-29533800-fhgsq\" (UID: \"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.310545 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g8bc\" (UniqueName: \"kubernetes.io/projected/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-kube-api-access-9g8bc\") pod \"collect-profiles-29533800-fhgsq\" (UID: \"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.310583 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-secret-volume\") pod \"collect-profiles-29533800-fhgsq\" (UID: \"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.310637 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwflf\" (UniqueName: \"kubernetes.io/projected/46c73af1-a231-49fc-aa2e-0278b7490eb1-kube-api-access-kwflf\") pod \"auto-csr-approver-29533800-6drfm\" (UID: \"46c73af1-a231-49fc-aa2e-0278b7490eb1\") " pod="openshift-infra/auto-csr-approver-29533800-6drfm" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.413438 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-config-volume\") pod \"collect-profiles-29533800-fhgsq\" (UID: \"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.413842 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g8bc\" (UniqueName: \"kubernetes.io/projected/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-kube-api-access-9g8bc\") pod \"collect-profiles-29533800-fhgsq\" (UID: \"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.413975 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-secret-volume\") pod \"collect-profiles-29533800-fhgsq\" (UID: \"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.414170 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwflf\" (UniqueName: \"kubernetes.io/projected/46c73af1-a231-49fc-aa2e-0278b7490eb1-kube-api-access-kwflf\") pod \"auto-csr-approver-29533800-6drfm\" (UID: \"46c73af1-a231-49fc-aa2e-0278b7490eb1\") " pod="openshift-infra/auto-csr-approver-29533800-6drfm" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.414477 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-config-volume\") pod \"collect-profiles-29533800-fhgsq\" (UID: \"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.420903 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-secret-volume\") pod \"collect-profiles-29533800-fhgsq\" (UID: \"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.443748 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g8bc\" (UniqueName: \"kubernetes.io/projected/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-kube-api-access-9g8bc\") pod \"collect-profiles-29533800-fhgsq\" (UID: \"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.449300 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwflf\" (UniqueName: \"kubernetes.io/projected/46c73af1-a231-49fc-aa2e-0278b7490eb1-kube-api-access-kwflf\") pod \"auto-csr-approver-29533800-6drfm\" (UID: \"46c73af1-a231-49fc-aa2e-0278b7490eb1\") " pod="openshift-infra/auto-csr-approver-29533800-6drfm" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.492798 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.514334 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533800-6drfm" Feb 25 14:00:00 crc kubenswrapper[4815]: I0225 14:00:00.974692 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533800-6drfm"] Feb 25 14:00:01 crc kubenswrapper[4815]: I0225 14:00:01.031122 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq"] Feb 25 14:00:01 crc kubenswrapper[4815]: E0225 14:00:01.484649 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13ae2bf8_e816_4d2f_9f9b_bb2a0bcb6e73.slice/crio-1fce3c81ab26c8eb8ed8f6422ebcfeffdb02f9dc3e94a87eb7e2c0fcd4806d9c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13ae2bf8_e816_4d2f_9f9b_bb2a0bcb6e73.slice/crio-conmon-1fce3c81ab26c8eb8ed8f6422ebcfeffdb02f9dc3e94a87eb7e2c0fcd4806d9c.scope\": RecentStats: unable to find data in memory cache]" Feb 25 14:00:01 crc kubenswrapper[4815]: I0225 14:00:01.772728 4815 generic.go:334] "Generic (PLEG): container finished" podID="13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73" containerID="1fce3c81ab26c8eb8ed8f6422ebcfeffdb02f9dc3e94a87eb7e2c0fcd4806d9c" exitCode=0 Feb 25 14:00:01 crc kubenswrapper[4815]: I0225 14:00:01.772810 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq" event={"ID":"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73","Type":"ContainerDied","Data":"1fce3c81ab26c8eb8ed8f6422ebcfeffdb02f9dc3e94a87eb7e2c0fcd4806d9c"} Feb 25 14:00:01 crc kubenswrapper[4815]: I0225 14:00:01.772844 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq" event={"ID":"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73","Type":"ContainerStarted","Data":"ceb843a5f69c35da6891b3dfb1cd86f8a2fd1ec8e446af11f79e4668c3effb71"} Feb 25 14:00:01 crc kubenswrapper[4815]: I0225 14:00:01.774371 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533800-6drfm" event={"ID":"46c73af1-a231-49fc-aa2e-0278b7490eb1","Type":"ContainerStarted","Data":"274d5fd221f08e9a9cdba7669a19381d82ca441996b920bf7a0fb239a2274202"} Feb 25 14:00:03 crc kubenswrapper[4815]: I0225 14:00:03.472695 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq" Feb 25 14:00:03 crc kubenswrapper[4815]: I0225 14:00:03.618049 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-config-volume\") pod \"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73\" (UID: \"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73\") " Feb 25 14:00:03 crc kubenswrapper[4815]: I0225 14:00:03.618419 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g8bc\" (UniqueName: \"kubernetes.io/projected/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-kube-api-access-9g8bc\") pod \"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73\" (UID: \"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73\") " Feb 25 14:00:03 crc kubenswrapper[4815]: I0225 14:00:03.618584 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-secret-volume\") pod \"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73\" (UID: \"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73\") " Feb 25 14:00:03 crc kubenswrapper[4815]: I0225 14:00:03.618758 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-config-volume" (OuterVolumeSpecName: "config-volume") pod "13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73" (UID: "13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 14:00:03 crc kubenswrapper[4815]: I0225 14:00:03.618963 4815 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 14:00:03 crc kubenswrapper[4815]: I0225 14:00:03.625217 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-kube-api-access-9g8bc" (OuterVolumeSpecName: "kube-api-access-9g8bc") pod "13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73" (UID: "13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73"). InnerVolumeSpecName "kube-api-access-9g8bc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:00:03 crc kubenswrapper[4815]: I0225 14:00:03.625399 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73" (UID: "13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 14:00:03 crc kubenswrapper[4815]: I0225 14:00:03.721016 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9g8bc\" (UniqueName: \"kubernetes.io/projected/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-kube-api-access-9g8bc\") on node \"crc\" DevicePath \"\"" Feb 25 14:00:03 crc kubenswrapper[4815]: I0225 14:00:03.721088 4815 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 14:00:03 crc kubenswrapper[4815]: I0225 14:00:03.794010 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq" event={"ID":"13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73","Type":"ContainerDied","Data":"ceb843a5f69c35da6891b3dfb1cd86f8a2fd1ec8e446af11f79e4668c3effb71"} Feb 25 14:00:03 crc kubenswrapper[4815]: I0225 14:00:03.794097 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ceb843a5f69c35da6891b3dfb1cd86f8a2fd1ec8e446af11f79e4668c3effb71" Feb 25 14:00:03 crc kubenswrapper[4815]: I0225 14:00:03.794052 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533800-fhgsq" Feb 25 14:00:03 crc kubenswrapper[4815]: I0225 14:00:03.936063 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 14:00:03 crc kubenswrapper[4815]: E0225 14:00:03.936466 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:00:04 crc kubenswrapper[4815]: I0225 14:00:04.545647 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv"] Feb 25 14:00:04 crc kubenswrapper[4815]: I0225 14:00:04.553751 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533755-65vjv"] Feb 25 14:00:04 crc kubenswrapper[4815]: I0225 14:00:04.947900 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80629dad-7d0a-46b3-bb52-09b75e7dfeb9" path="/var/lib/kubelet/pods/80629dad-7d0a-46b3-bb52-09b75e7dfeb9/volumes" Feb 25 14:00:05 crc kubenswrapper[4815]: I0225 14:00:05.811794 4815 generic.go:334] "Generic (PLEG): container finished" podID="46c73af1-a231-49fc-aa2e-0278b7490eb1" containerID="6ca4da694e28152b3541b51daf3321e9636dddf07bee483b236ebb5c0509eadb" exitCode=0 Feb 25 14:00:05 crc kubenswrapper[4815]: I0225 14:00:05.811863 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533800-6drfm" event={"ID":"46c73af1-a231-49fc-aa2e-0278b7490eb1","Type":"ContainerDied","Data":"6ca4da694e28152b3541b51daf3321e9636dddf07bee483b236ebb5c0509eadb"} Feb 25 14:00:07 crc kubenswrapper[4815]: I0225 14:00:07.231799 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533800-6drfm" Feb 25 14:00:07 crc kubenswrapper[4815]: I0225 14:00:07.390249 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwflf\" (UniqueName: \"kubernetes.io/projected/46c73af1-a231-49fc-aa2e-0278b7490eb1-kube-api-access-kwflf\") pod \"46c73af1-a231-49fc-aa2e-0278b7490eb1\" (UID: \"46c73af1-a231-49fc-aa2e-0278b7490eb1\") " Feb 25 14:00:07 crc kubenswrapper[4815]: I0225 14:00:07.401214 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46c73af1-a231-49fc-aa2e-0278b7490eb1-kube-api-access-kwflf" (OuterVolumeSpecName: "kube-api-access-kwflf") pod "46c73af1-a231-49fc-aa2e-0278b7490eb1" (UID: "46c73af1-a231-49fc-aa2e-0278b7490eb1"). InnerVolumeSpecName "kube-api-access-kwflf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:00:07 crc kubenswrapper[4815]: I0225 14:00:07.494455 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwflf\" (UniqueName: \"kubernetes.io/projected/46c73af1-a231-49fc-aa2e-0278b7490eb1-kube-api-access-kwflf\") on node \"crc\" DevicePath \"\"" Feb 25 14:00:07 crc kubenswrapper[4815]: I0225 14:00:07.832308 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533800-6drfm" event={"ID":"46c73af1-a231-49fc-aa2e-0278b7490eb1","Type":"ContainerDied","Data":"274d5fd221f08e9a9cdba7669a19381d82ca441996b920bf7a0fb239a2274202"} Feb 25 14:00:07 crc kubenswrapper[4815]: I0225 14:00:07.832544 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="274d5fd221f08e9a9cdba7669a19381d82ca441996b920bf7a0fb239a2274202" Feb 25 14:00:07 crc kubenswrapper[4815]: I0225 14:00:07.832377 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533800-6drfm" Feb 25 14:00:08 crc kubenswrapper[4815]: I0225 14:00:08.293015 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533794-l8z6b"] Feb 25 14:00:08 crc kubenswrapper[4815]: I0225 14:00:08.305219 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533794-l8z6b"] Feb 25 14:00:08 crc kubenswrapper[4815]: I0225 14:00:08.953043 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc77125d-9f98-41b8-a366-d35124c45f9f" path="/var/lib/kubelet/pods/fc77125d-9f98-41b8-a366-d35124c45f9f/volumes" Feb 25 14:00:18 crc kubenswrapper[4815]: I0225 14:00:18.935668 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 14:00:18 crc kubenswrapper[4815]: E0225 14:00:18.936601 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:00:30 crc kubenswrapper[4815]: I0225 14:00:30.935625 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 14:00:30 crc kubenswrapper[4815]: E0225 14:00:30.936683 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:00:45 crc kubenswrapper[4815]: I0225 14:00:45.935636 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 14:00:45 crc kubenswrapper[4815]: E0225 14:00:45.936405 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:00:48 crc kubenswrapper[4815]: I0225 14:00:48.658096 4815 scope.go:117] "RemoveContainer" containerID="ea3d582b548bf242a266e784b53beb6b69a0e08bfae98fb3536a7c6fed44fe63" Feb 25 14:00:48 crc kubenswrapper[4815]: I0225 14:00:48.703569 4815 scope.go:117] "RemoveContainer" containerID="5fcc6127de7c4f26b8ab72ffdab14f84fee434383c588d93cc0b01819e27c9cb" Feb 25 14:00:49 crc kubenswrapper[4815]: I0225 14:00:49.275145 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wqc5f"] Feb 25 14:00:49 crc kubenswrapper[4815]: E0225 14:00:49.275989 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73" containerName="collect-profiles" Feb 25 14:00:49 crc kubenswrapper[4815]: I0225 14:00:49.276019 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73" containerName="collect-profiles" Feb 25 14:00:49 crc kubenswrapper[4815]: E0225 14:00:49.276034 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46c73af1-a231-49fc-aa2e-0278b7490eb1" containerName="oc" Feb 25 14:00:49 crc kubenswrapper[4815]: I0225 14:00:49.276042 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="46c73af1-a231-49fc-aa2e-0278b7490eb1" containerName="oc" Feb 25 14:00:49 crc kubenswrapper[4815]: I0225 14:00:49.276271 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="13ae2bf8-e816-4d2f-9f9b-bb2a0bcb6e73" containerName="collect-profiles" Feb 25 14:00:49 crc kubenswrapper[4815]: I0225 14:00:49.276301 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="46c73af1-a231-49fc-aa2e-0278b7490eb1" containerName="oc" Feb 25 14:00:49 crc kubenswrapper[4815]: I0225 14:00:49.278030 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqc5f" Feb 25 14:00:49 crc kubenswrapper[4815]: I0225 14:00:49.294019 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wqc5f"] Feb 25 14:00:49 crc kubenswrapper[4815]: I0225 14:00:49.439023 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c12242f-0da5-4f59-84e7-77bb094a57dc-utilities\") pod \"certified-operators-wqc5f\" (UID: \"7c12242f-0da5-4f59-84e7-77bb094a57dc\") " pod="openshift-marketplace/certified-operators-wqc5f" Feb 25 14:00:49 crc kubenswrapper[4815]: I0225 14:00:49.439097 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c12242f-0da5-4f59-84e7-77bb094a57dc-catalog-content\") pod \"certified-operators-wqc5f\" (UID: \"7c12242f-0da5-4f59-84e7-77bb094a57dc\") " pod="openshift-marketplace/certified-operators-wqc5f" Feb 25 14:00:49 crc kubenswrapper[4815]: I0225 14:00:49.439394 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94jl4\" (UniqueName: \"kubernetes.io/projected/7c12242f-0da5-4f59-84e7-77bb094a57dc-kube-api-access-94jl4\") pod \"certified-operators-wqc5f\" (UID: \"7c12242f-0da5-4f59-84e7-77bb094a57dc\") " pod="openshift-marketplace/certified-operators-wqc5f" Feb 25 14:00:49 crc kubenswrapper[4815]: I0225 14:00:49.540916 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94jl4\" (UniqueName: \"kubernetes.io/projected/7c12242f-0da5-4f59-84e7-77bb094a57dc-kube-api-access-94jl4\") pod \"certified-operators-wqc5f\" (UID: \"7c12242f-0da5-4f59-84e7-77bb094a57dc\") " pod="openshift-marketplace/certified-operators-wqc5f" Feb 25 14:00:49 crc kubenswrapper[4815]: I0225 14:00:49.541491 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c12242f-0da5-4f59-84e7-77bb094a57dc-utilities\") pod \"certified-operators-wqc5f\" (UID: \"7c12242f-0da5-4f59-84e7-77bb094a57dc\") " pod="openshift-marketplace/certified-operators-wqc5f" Feb 25 14:00:49 crc kubenswrapper[4815]: I0225 14:00:49.541711 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c12242f-0da5-4f59-84e7-77bb094a57dc-catalog-content\") pod \"certified-operators-wqc5f\" (UID: \"7c12242f-0da5-4f59-84e7-77bb094a57dc\") " pod="openshift-marketplace/certified-operators-wqc5f" Feb 25 14:00:49 crc kubenswrapper[4815]: I0225 14:00:49.542096 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c12242f-0da5-4f59-84e7-77bb094a57dc-utilities\") pod \"certified-operators-wqc5f\" (UID: \"7c12242f-0da5-4f59-84e7-77bb094a57dc\") " pod="openshift-marketplace/certified-operators-wqc5f" Feb 25 14:00:49 crc kubenswrapper[4815]: I0225 14:00:49.542130 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c12242f-0da5-4f59-84e7-77bb094a57dc-catalog-content\") pod \"certified-operators-wqc5f\" (UID: \"7c12242f-0da5-4f59-84e7-77bb094a57dc\") " pod="openshift-marketplace/certified-operators-wqc5f" Feb 25 14:00:49 crc kubenswrapper[4815]: I0225 14:00:49.562286 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94jl4\" (UniqueName: \"kubernetes.io/projected/7c12242f-0da5-4f59-84e7-77bb094a57dc-kube-api-access-94jl4\") pod \"certified-operators-wqc5f\" (UID: \"7c12242f-0da5-4f59-84e7-77bb094a57dc\") " pod="openshift-marketplace/certified-operators-wqc5f" Feb 25 14:00:49 crc kubenswrapper[4815]: I0225 14:00:49.622770 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqc5f" Feb 25 14:00:50 crc kubenswrapper[4815]: I0225 14:00:50.143468 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wqc5f"] Feb 25 14:00:50 crc kubenswrapper[4815]: I0225 14:00:50.257707 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqc5f" event={"ID":"7c12242f-0da5-4f59-84e7-77bb094a57dc","Type":"ContainerStarted","Data":"4453b3a2a000e5c58a955e0ba99ed725fadc3036971dab9410776bc6a1d9cbb2"} Feb 25 14:00:50 crc kubenswrapper[4815]: I0225 14:00:50.676100 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zk26m"] Feb 25 14:00:50 crc kubenswrapper[4815]: I0225 14:00:50.678436 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zk26m" Feb 25 14:00:50 crc kubenswrapper[4815]: I0225 14:00:50.688693 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zk26m"] Feb 25 14:00:50 crc kubenswrapper[4815]: I0225 14:00:50.766562 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-utilities\") pod \"community-operators-zk26m\" (UID: \"12c4d0d5-cdf5-4968-9fde-abe8fac843bd\") " pod="openshift-marketplace/community-operators-zk26m" Feb 25 14:00:50 crc kubenswrapper[4815]: I0225 14:00:50.766838 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjwhw\" (UniqueName: \"kubernetes.io/projected/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-kube-api-access-jjwhw\") pod \"community-operators-zk26m\" (UID: \"12c4d0d5-cdf5-4968-9fde-abe8fac843bd\") " pod="openshift-marketplace/community-operators-zk26m" Feb 25 14:00:50 crc kubenswrapper[4815]: I0225 14:00:50.767308 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-catalog-content\") pod \"community-operators-zk26m\" (UID: \"12c4d0d5-cdf5-4968-9fde-abe8fac843bd\") " pod="openshift-marketplace/community-operators-zk26m" Feb 25 14:00:50 crc kubenswrapper[4815]: I0225 14:00:50.869598 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-catalog-content\") pod \"community-operators-zk26m\" (UID: \"12c4d0d5-cdf5-4968-9fde-abe8fac843bd\") " pod="openshift-marketplace/community-operators-zk26m" Feb 25 14:00:50 crc kubenswrapper[4815]: I0225 14:00:50.869642 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-utilities\") pod \"community-operators-zk26m\" (UID: \"12c4d0d5-cdf5-4968-9fde-abe8fac843bd\") " pod="openshift-marketplace/community-operators-zk26m" Feb 25 14:00:50 crc kubenswrapper[4815]: I0225 14:00:50.869703 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjwhw\" (UniqueName: \"kubernetes.io/projected/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-kube-api-access-jjwhw\") pod \"community-operators-zk26m\" (UID: \"12c4d0d5-cdf5-4968-9fde-abe8fac843bd\") " pod="openshift-marketplace/community-operators-zk26m" Feb 25 14:00:50 crc kubenswrapper[4815]: I0225 14:00:50.870290 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-catalog-content\") pod \"community-operators-zk26m\" (UID: \"12c4d0d5-cdf5-4968-9fde-abe8fac843bd\") " pod="openshift-marketplace/community-operators-zk26m" Feb 25 14:00:50 crc kubenswrapper[4815]: I0225 14:00:50.870453 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-utilities\") pod \"community-operators-zk26m\" (UID: \"12c4d0d5-cdf5-4968-9fde-abe8fac843bd\") " pod="openshift-marketplace/community-operators-zk26m" Feb 25 14:00:50 crc kubenswrapper[4815]: I0225 14:00:50.896785 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjwhw\" (UniqueName: \"kubernetes.io/projected/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-kube-api-access-jjwhw\") pod \"community-operators-zk26m\" (UID: \"12c4d0d5-cdf5-4968-9fde-abe8fac843bd\") " pod="openshift-marketplace/community-operators-zk26m" Feb 25 14:00:50 crc kubenswrapper[4815]: I0225 14:00:50.994084 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zk26m" Feb 25 14:00:51 crc kubenswrapper[4815]: I0225 14:00:51.279040 4815 generic.go:334] "Generic (PLEG): container finished" podID="7c12242f-0da5-4f59-84e7-77bb094a57dc" containerID="b90f4e2fb5d0ab99a7243fbfffd3eb626f501903334558c6c484b9ec78b25f3a" exitCode=0 Feb 25 14:00:51 crc kubenswrapper[4815]: I0225 14:00:51.279087 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqc5f" event={"ID":"7c12242f-0da5-4f59-84e7-77bb094a57dc","Type":"ContainerDied","Data":"b90f4e2fb5d0ab99a7243fbfffd3eb626f501903334558c6c484b9ec78b25f3a"} Feb 25 14:00:51 crc kubenswrapper[4815]: I0225 14:00:51.283832 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 14:00:51 crc kubenswrapper[4815]: I0225 14:00:51.493623 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zk26m"] Feb 25 14:00:51 crc kubenswrapper[4815]: W0225 14:00:51.503930 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12c4d0d5_cdf5_4968_9fde_abe8fac843bd.slice/crio-7f618282630517c0f02566230ecca83d3221e8aec4039bec2a1922284d8e52b1 WatchSource:0}: Error finding container 7f618282630517c0f02566230ecca83d3221e8aec4039bec2a1922284d8e52b1: Status 404 returned error can't find the container with id 7f618282630517c0f02566230ecca83d3221e8aec4039bec2a1922284d8e52b1 Feb 25 14:00:52 crc kubenswrapper[4815]: I0225 14:00:52.291148 4815 generic.go:334] "Generic (PLEG): container finished" podID="12c4d0d5-cdf5-4968-9fde-abe8fac843bd" containerID="1906026ad6b21944ed5ec2807d97d4279238deafa5963c99e427e9e0d5ba0459" exitCode=0 Feb 25 14:00:52 crc kubenswrapper[4815]: I0225 14:00:52.291238 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk26m" event={"ID":"12c4d0d5-cdf5-4968-9fde-abe8fac843bd","Type":"ContainerDied","Data":"1906026ad6b21944ed5ec2807d97d4279238deafa5963c99e427e9e0d5ba0459"} Feb 25 14:00:52 crc kubenswrapper[4815]: I0225 14:00:52.291559 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk26m" event={"ID":"12c4d0d5-cdf5-4968-9fde-abe8fac843bd","Type":"ContainerStarted","Data":"7f618282630517c0f02566230ecca83d3221e8aec4039bec2a1922284d8e52b1"} Feb 25 14:00:52 crc kubenswrapper[4815]: I0225 14:00:52.294103 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqc5f" event={"ID":"7c12242f-0da5-4f59-84e7-77bb094a57dc","Type":"ContainerStarted","Data":"7a748c4f9df5ebfec532dd4d918f662e48fdb4322332f811fc839d1db42383b6"} Feb 25 14:00:53 crc kubenswrapper[4815]: I0225 14:00:53.307751 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk26m" event={"ID":"12c4d0d5-cdf5-4968-9fde-abe8fac843bd","Type":"ContainerStarted","Data":"e4e6980f17db689d40cebc70e764672f0406e64b46c28ed0b4832d906c1364be"} Feb 25 14:00:54 crc kubenswrapper[4815]: I0225 14:00:54.322770 4815 generic.go:334] "Generic (PLEG): container finished" podID="12c4d0d5-cdf5-4968-9fde-abe8fac843bd" containerID="e4e6980f17db689d40cebc70e764672f0406e64b46c28ed0b4832d906c1364be" exitCode=0 Feb 25 14:00:54 crc kubenswrapper[4815]: I0225 14:00:54.322855 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk26m" event={"ID":"12c4d0d5-cdf5-4968-9fde-abe8fac843bd","Type":"ContainerDied","Data":"e4e6980f17db689d40cebc70e764672f0406e64b46c28ed0b4832d906c1364be"} Feb 25 14:00:54 crc kubenswrapper[4815]: I0225 14:00:54.328103 4815 generic.go:334] "Generic (PLEG): container finished" podID="7c12242f-0da5-4f59-84e7-77bb094a57dc" containerID="7a748c4f9df5ebfec532dd4d918f662e48fdb4322332f811fc839d1db42383b6" exitCode=0 Feb 25 14:00:54 crc kubenswrapper[4815]: I0225 14:00:54.328147 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqc5f" event={"ID":"7c12242f-0da5-4f59-84e7-77bb094a57dc","Type":"ContainerDied","Data":"7a748c4f9df5ebfec532dd4d918f662e48fdb4322332f811fc839d1db42383b6"} Feb 25 14:00:55 crc kubenswrapper[4815]: I0225 14:00:55.356213 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk26m" event={"ID":"12c4d0d5-cdf5-4968-9fde-abe8fac843bd","Type":"ContainerStarted","Data":"36ee872a62d8cc4816c2515fea65f812987830a97f63adeff30e3dba7fcc58d3"} Feb 25 14:00:55 crc kubenswrapper[4815]: I0225 14:00:55.362907 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqc5f" event={"ID":"7c12242f-0da5-4f59-84e7-77bb094a57dc","Type":"ContainerStarted","Data":"da4230c4b4d25b5e6815aaea5ee66ca3f093224f943a39f486ee317460459dab"} Feb 25 14:00:55 crc kubenswrapper[4815]: I0225 14:00:55.380858 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zk26m" podStartSLOduration=2.893996284 podStartE2EDuration="5.380836819s" podCreationTimestamp="2026-02-25 14:00:50 +0000 UTC" firstStartedPulling="2026-02-25 14:00:52.295695317 +0000 UTC m=+2410.096793381" lastFinishedPulling="2026-02-25 14:00:54.782535862 +0000 UTC m=+2412.583633916" observedRunningTime="2026-02-25 14:00:55.375571815 +0000 UTC m=+2413.176669889" watchObservedRunningTime="2026-02-25 14:00:55.380836819 +0000 UTC m=+2413.181934873" Feb 25 14:00:55 crc kubenswrapper[4815]: I0225 14:00:55.400670 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wqc5f" podStartSLOduration=2.978938629 podStartE2EDuration="6.400651809s" podCreationTimestamp="2026-02-25 14:00:49 +0000 UTC" firstStartedPulling="2026-02-25 14:00:51.283552908 +0000 UTC m=+2409.084650952" lastFinishedPulling="2026-02-25 14:00:54.705266068 +0000 UTC m=+2412.506364132" observedRunningTime="2026-02-25 14:00:55.393532406 +0000 UTC m=+2413.194630480" watchObservedRunningTime="2026-02-25 14:00:55.400651809 +0000 UTC m=+2413.201749863" Feb 25 14:00:59 crc kubenswrapper[4815]: I0225 14:00:59.623823 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wqc5f" Feb 25 14:00:59 crc kubenswrapper[4815]: I0225 14:00:59.624401 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wqc5f" Feb 25 14:00:59 crc kubenswrapper[4815]: I0225 14:00:59.673738 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wqc5f" Feb 25 14:00:59 crc kubenswrapper[4815]: I0225 14:00:59.935902 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 14:00:59 crc kubenswrapper[4815]: E0225 14:00:59.936247 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.162001 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29533801-mmp7n"] Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.163975 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29533801-mmp7n" Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.177131 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29533801-mmp7n"] Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.251390 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-combined-ca-bundle\") pod \"keystone-cron-29533801-mmp7n\" (UID: \"99ed322b-3f0d-4adc-8244-07af594b14e7\") " pod="openstack/keystone-cron-29533801-mmp7n" Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.251442 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-fernet-keys\") pod \"keystone-cron-29533801-mmp7n\" (UID: \"99ed322b-3f0d-4adc-8244-07af594b14e7\") " pod="openstack/keystone-cron-29533801-mmp7n" Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.251496 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-config-data\") pod \"keystone-cron-29533801-mmp7n\" (UID: \"99ed322b-3f0d-4adc-8244-07af594b14e7\") " pod="openstack/keystone-cron-29533801-mmp7n" Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.251547 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmkjt\" (UniqueName: \"kubernetes.io/projected/99ed322b-3f0d-4adc-8244-07af594b14e7-kube-api-access-vmkjt\") pod \"keystone-cron-29533801-mmp7n\" (UID: \"99ed322b-3f0d-4adc-8244-07af594b14e7\") " pod="openstack/keystone-cron-29533801-mmp7n" Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.353026 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-combined-ca-bundle\") pod \"keystone-cron-29533801-mmp7n\" (UID: \"99ed322b-3f0d-4adc-8244-07af594b14e7\") " pod="openstack/keystone-cron-29533801-mmp7n" Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.353271 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-fernet-keys\") pod \"keystone-cron-29533801-mmp7n\" (UID: \"99ed322b-3f0d-4adc-8244-07af594b14e7\") " pod="openstack/keystone-cron-29533801-mmp7n" Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.353408 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-config-data\") pod \"keystone-cron-29533801-mmp7n\" (UID: \"99ed322b-3f0d-4adc-8244-07af594b14e7\") " pod="openstack/keystone-cron-29533801-mmp7n" Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.353539 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmkjt\" (UniqueName: \"kubernetes.io/projected/99ed322b-3f0d-4adc-8244-07af594b14e7-kube-api-access-vmkjt\") pod \"keystone-cron-29533801-mmp7n\" (UID: \"99ed322b-3f0d-4adc-8244-07af594b14e7\") " pod="openstack/keystone-cron-29533801-mmp7n" Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.359045 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-combined-ca-bundle\") pod \"keystone-cron-29533801-mmp7n\" (UID: \"99ed322b-3f0d-4adc-8244-07af594b14e7\") " pod="openstack/keystone-cron-29533801-mmp7n" Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.359999 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-config-data\") pod \"keystone-cron-29533801-mmp7n\" (UID: \"99ed322b-3f0d-4adc-8244-07af594b14e7\") " pod="openstack/keystone-cron-29533801-mmp7n" Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.360322 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-fernet-keys\") pod \"keystone-cron-29533801-mmp7n\" (UID: \"99ed322b-3f0d-4adc-8244-07af594b14e7\") " pod="openstack/keystone-cron-29533801-mmp7n" Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.371643 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmkjt\" (UniqueName: \"kubernetes.io/projected/99ed322b-3f0d-4adc-8244-07af594b14e7-kube-api-access-vmkjt\") pod \"keystone-cron-29533801-mmp7n\" (UID: \"99ed322b-3f0d-4adc-8244-07af594b14e7\") " pod="openstack/keystone-cron-29533801-mmp7n" Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.459651 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wqc5f" Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.505611 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29533801-mmp7n" Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.947182 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29533801-mmp7n"] Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.995024 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zk26m" Feb 25 14:01:00 crc kubenswrapper[4815]: I0225 14:01:00.995072 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zk26m" Feb 25 14:01:01 crc kubenswrapper[4815]: I0225 14:01:01.041787 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zk26m" Feb 25 14:01:01 crc kubenswrapper[4815]: I0225 14:01:01.421650 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29533801-mmp7n" event={"ID":"99ed322b-3f0d-4adc-8244-07af594b14e7","Type":"ContainerStarted","Data":"20bc704cc1379f391124b722b7d784adf09fd3f67054b1b8d662797b6da82545"} Feb 25 14:01:01 crc kubenswrapper[4815]: I0225 14:01:01.421937 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29533801-mmp7n" event={"ID":"99ed322b-3f0d-4adc-8244-07af594b14e7","Type":"ContainerStarted","Data":"074ee107c659618b8900723aba69ea68879a8cc41f8db3c920b1eb6114ae9653"} Feb 25 14:01:01 crc kubenswrapper[4815]: I0225 14:01:01.450616 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29533801-mmp7n" podStartSLOduration=1.450594201 podStartE2EDuration="1.450594201s" podCreationTimestamp="2026-02-25 14:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 14:01:01.439698711 +0000 UTC m=+2419.240796765" watchObservedRunningTime="2026-02-25 14:01:01.450594201 +0000 UTC m=+2419.251692255" Feb 25 14:01:01 crc kubenswrapper[4815]: I0225 14:01:01.483716 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zk26m" Feb 25 14:01:03 crc kubenswrapper[4815]: I0225 14:01:03.453966 4815 generic.go:334] "Generic (PLEG): container finished" podID="99ed322b-3f0d-4adc-8244-07af594b14e7" containerID="20bc704cc1379f391124b722b7d784adf09fd3f67054b1b8d662797b6da82545" exitCode=0 Feb 25 14:01:03 crc kubenswrapper[4815]: I0225 14:01:03.454055 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29533801-mmp7n" event={"ID":"99ed322b-3f0d-4adc-8244-07af594b14e7","Type":"ContainerDied","Data":"20bc704cc1379f391124b722b7d784adf09fd3f67054b1b8d662797b6da82545"} Feb 25 14:01:04 crc kubenswrapper[4815]: I0225 14:01:04.453893 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wqc5f"] Feb 25 14:01:04 crc kubenswrapper[4815]: I0225 14:01:04.454414 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wqc5f" podUID="7c12242f-0da5-4f59-84e7-77bb094a57dc" containerName="registry-server" containerID="cri-o://da4230c4b4d25b5e6815aaea5ee66ca3f093224f943a39f486ee317460459dab" gracePeriod=2 Feb 25 14:01:04 crc kubenswrapper[4815]: I0225 14:01:04.652929 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zk26m"] Feb 25 14:01:04 crc kubenswrapper[4815]: I0225 14:01:04.653170 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zk26m" podUID="12c4d0d5-cdf5-4968-9fde-abe8fac843bd" containerName="registry-server" containerID="cri-o://36ee872a62d8cc4816c2515fea65f812987830a97f63adeff30e3dba7fcc58d3" gracePeriod=2 Feb 25 14:01:04 crc kubenswrapper[4815]: I0225 14:01:04.834121 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29533801-mmp7n" Feb 25 14:01:04 crc kubenswrapper[4815]: I0225 14:01:04.940001 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqc5f" Feb 25 14:01:04 crc kubenswrapper[4815]: I0225 14:01:04.944537 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-fernet-keys\") pod \"99ed322b-3f0d-4adc-8244-07af594b14e7\" (UID: \"99ed322b-3f0d-4adc-8244-07af594b14e7\") " Feb 25 14:01:04 crc kubenswrapper[4815]: I0225 14:01:04.945883 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmkjt\" (UniqueName: \"kubernetes.io/projected/99ed322b-3f0d-4adc-8244-07af594b14e7-kube-api-access-vmkjt\") pod \"99ed322b-3f0d-4adc-8244-07af594b14e7\" (UID: \"99ed322b-3f0d-4adc-8244-07af594b14e7\") " Feb 25 14:01:04 crc kubenswrapper[4815]: I0225 14:01:04.945990 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-combined-ca-bundle\") pod \"99ed322b-3f0d-4adc-8244-07af594b14e7\" (UID: \"99ed322b-3f0d-4adc-8244-07af594b14e7\") " Feb 25 14:01:04 crc kubenswrapper[4815]: I0225 14:01:04.946020 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-config-data\") pod \"99ed322b-3f0d-4adc-8244-07af594b14e7\" (UID: \"99ed322b-3f0d-4adc-8244-07af594b14e7\") " Feb 25 14:01:04 crc kubenswrapper[4815]: I0225 14:01:04.953157 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "99ed322b-3f0d-4adc-8244-07af594b14e7" (UID: "99ed322b-3f0d-4adc-8244-07af594b14e7"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 14:01:04 crc kubenswrapper[4815]: I0225 14:01:04.953174 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99ed322b-3f0d-4adc-8244-07af594b14e7-kube-api-access-vmkjt" (OuterVolumeSpecName: "kube-api-access-vmkjt") pod "99ed322b-3f0d-4adc-8244-07af594b14e7" (UID: "99ed322b-3f0d-4adc-8244-07af594b14e7"). InnerVolumeSpecName "kube-api-access-vmkjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:04.990484 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99ed322b-3f0d-4adc-8244-07af594b14e7" (UID: "99ed322b-3f0d-4adc-8244-07af594b14e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.018495 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-config-data" (OuterVolumeSpecName: "config-data") pod "99ed322b-3f0d-4adc-8244-07af594b14e7" (UID: "99ed322b-3f0d-4adc-8244-07af594b14e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.048439 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c12242f-0da5-4f59-84e7-77bb094a57dc-utilities\") pod \"7c12242f-0da5-4f59-84e7-77bb094a57dc\" (UID: \"7c12242f-0da5-4f59-84e7-77bb094a57dc\") " Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.048583 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c12242f-0da5-4f59-84e7-77bb094a57dc-catalog-content\") pod \"7c12242f-0da5-4f59-84e7-77bb094a57dc\" (UID: \"7c12242f-0da5-4f59-84e7-77bb094a57dc\") " Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.048666 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94jl4\" (UniqueName: \"kubernetes.io/projected/7c12242f-0da5-4f59-84e7-77bb094a57dc-kube-api-access-94jl4\") pod \"7c12242f-0da5-4f59-84e7-77bb094a57dc\" (UID: \"7c12242f-0da5-4f59-84e7-77bb094a57dc\") " Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.049088 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmkjt\" (UniqueName: \"kubernetes.io/projected/99ed322b-3f0d-4adc-8244-07af594b14e7-kube-api-access-vmkjt\") on node \"crc\" DevicePath \"\"" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.049104 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.049118 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.049129 4815 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/99ed322b-3f0d-4adc-8244-07af594b14e7-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.049423 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c12242f-0da5-4f59-84e7-77bb094a57dc-utilities" (OuterVolumeSpecName: "utilities") pod "7c12242f-0da5-4f59-84e7-77bb094a57dc" (UID: "7c12242f-0da5-4f59-84e7-77bb094a57dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.051955 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c12242f-0da5-4f59-84e7-77bb094a57dc-kube-api-access-94jl4" (OuterVolumeSpecName: "kube-api-access-94jl4") pod "7c12242f-0da5-4f59-84e7-77bb094a57dc" (UID: "7c12242f-0da5-4f59-84e7-77bb094a57dc"). InnerVolumeSpecName "kube-api-access-94jl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.064233 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zk26m" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.115005 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c12242f-0da5-4f59-84e7-77bb094a57dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c12242f-0da5-4f59-84e7-77bb094a57dc" (UID: "7c12242f-0da5-4f59-84e7-77bb094a57dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.150716 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjwhw\" (UniqueName: \"kubernetes.io/projected/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-kube-api-access-jjwhw\") pod \"12c4d0d5-cdf5-4968-9fde-abe8fac843bd\" (UID: \"12c4d0d5-cdf5-4968-9fde-abe8fac843bd\") " Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.150962 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-utilities\") pod \"12c4d0d5-cdf5-4968-9fde-abe8fac843bd\" (UID: \"12c4d0d5-cdf5-4968-9fde-abe8fac843bd\") " Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.151008 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-catalog-content\") pod \"12c4d0d5-cdf5-4968-9fde-abe8fac843bd\" (UID: \"12c4d0d5-cdf5-4968-9fde-abe8fac843bd\") " Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.151416 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c12242f-0da5-4f59-84e7-77bb094a57dc-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.151429 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94jl4\" (UniqueName: \"kubernetes.io/projected/7c12242f-0da5-4f59-84e7-77bb094a57dc-kube-api-access-94jl4\") on node \"crc\" DevicePath \"\"" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.151440 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c12242f-0da5-4f59-84e7-77bb094a57dc-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.151798 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-utilities" (OuterVolumeSpecName: "utilities") pod "12c4d0d5-cdf5-4968-9fde-abe8fac843bd" (UID: "12c4d0d5-cdf5-4968-9fde-abe8fac843bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.168588 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-kube-api-access-jjwhw" (OuterVolumeSpecName: "kube-api-access-jjwhw") pod "12c4d0d5-cdf5-4968-9fde-abe8fac843bd" (UID: "12c4d0d5-cdf5-4968-9fde-abe8fac843bd"). InnerVolumeSpecName "kube-api-access-jjwhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.207403 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "12c4d0d5-cdf5-4968-9fde-abe8fac843bd" (UID: "12c4d0d5-cdf5-4968-9fde-abe8fac843bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.253413 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.253449 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.253464 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjwhw\" (UniqueName: \"kubernetes.io/projected/12c4d0d5-cdf5-4968-9fde-abe8fac843bd-kube-api-access-jjwhw\") on node \"crc\" DevicePath \"\"" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.487882 4815 generic.go:334] "Generic (PLEG): container finished" podID="12c4d0d5-cdf5-4968-9fde-abe8fac843bd" containerID="36ee872a62d8cc4816c2515fea65f812987830a97f63adeff30e3dba7fcc58d3" exitCode=0 Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.487956 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zk26m" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.487973 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk26m" event={"ID":"12c4d0d5-cdf5-4968-9fde-abe8fac843bd","Type":"ContainerDied","Data":"36ee872a62d8cc4816c2515fea65f812987830a97f63adeff30e3dba7fcc58d3"} Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.488015 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk26m" event={"ID":"12c4d0d5-cdf5-4968-9fde-abe8fac843bd","Type":"ContainerDied","Data":"7f618282630517c0f02566230ecca83d3221e8aec4039bec2a1922284d8e52b1"} Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.488041 4815 scope.go:117] "RemoveContainer" containerID="36ee872a62d8cc4816c2515fea65f812987830a97f63adeff30e3dba7fcc58d3" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.507317 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29533801-mmp7n" event={"ID":"99ed322b-3f0d-4adc-8244-07af594b14e7","Type":"ContainerDied","Data":"074ee107c659618b8900723aba69ea68879a8cc41f8db3c920b1eb6114ae9653"} Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.507355 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="074ee107c659618b8900723aba69ea68879a8cc41f8db3c920b1eb6114ae9653" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.507587 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29533801-mmp7n" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.517020 4815 generic.go:334] "Generic (PLEG): container finished" podID="7c12242f-0da5-4f59-84e7-77bb094a57dc" containerID="da4230c4b4d25b5e6815aaea5ee66ca3f093224f943a39f486ee317460459dab" exitCode=0 Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.517059 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqc5f" event={"ID":"7c12242f-0da5-4f59-84e7-77bb094a57dc","Type":"ContainerDied","Data":"da4230c4b4d25b5e6815aaea5ee66ca3f093224f943a39f486ee317460459dab"} Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.517084 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqc5f" event={"ID":"7c12242f-0da5-4f59-84e7-77bb094a57dc","Type":"ContainerDied","Data":"4453b3a2a000e5c58a955e0ba99ed725fadc3036971dab9410776bc6a1d9cbb2"} Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.517149 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqc5f" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.556716 4815 scope.go:117] "RemoveContainer" containerID="e4e6980f17db689d40cebc70e764672f0406e64b46c28ed0b4832d906c1364be" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.576649 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zk26m"] Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.588896 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zk26m"] Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.594194 4815 scope.go:117] "RemoveContainer" containerID="1906026ad6b21944ed5ec2807d97d4279238deafa5963c99e427e9e0d5ba0459" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.596685 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wqc5f"] Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.604132 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wqc5f"] Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.656581 4815 scope.go:117] "RemoveContainer" containerID="36ee872a62d8cc4816c2515fea65f812987830a97f63adeff30e3dba7fcc58d3" Feb 25 14:01:05 crc kubenswrapper[4815]: E0225 14:01:05.657060 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36ee872a62d8cc4816c2515fea65f812987830a97f63adeff30e3dba7fcc58d3\": container with ID starting with 36ee872a62d8cc4816c2515fea65f812987830a97f63adeff30e3dba7fcc58d3 not found: ID does not exist" containerID="36ee872a62d8cc4816c2515fea65f812987830a97f63adeff30e3dba7fcc58d3" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.657117 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36ee872a62d8cc4816c2515fea65f812987830a97f63adeff30e3dba7fcc58d3"} err="failed to get container status \"36ee872a62d8cc4816c2515fea65f812987830a97f63adeff30e3dba7fcc58d3\": rpc error: code = NotFound desc = could not find container \"36ee872a62d8cc4816c2515fea65f812987830a97f63adeff30e3dba7fcc58d3\": container with ID starting with 36ee872a62d8cc4816c2515fea65f812987830a97f63adeff30e3dba7fcc58d3 not found: ID does not exist" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.657154 4815 scope.go:117] "RemoveContainer" containerID="e4e6980f17db689d40cebc70e764672f0406e64b46c28ed0b4832d906c1364be" Feb 25 14:01:05 crc kubenswrapper[4815]: E0225 14:01:05.657471 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4e6980f17db689d40cebc70e764672f0406e64b46c28ed0b4832d906c1364be\": container with ID starting with e4e6980f17db689d40cebc70e764672f0406e64b46c28ed0b4832d906c1364be not found: ID does not exist" containerID="e4e6980f17db689d40cebc70e764672f0406e64b46c28ed0b4832d906c1364be" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.657500 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4e6980f17db689d40cebc70e764672f0406e64b46c28ed0b4832d906c1364be"} err="failed to get container status \"e4e6980f17db689d40cebc70e764672f0406e64b46c28ed0b4832d906c1364be\": rpc error: code = NotFound desc = could not find container \"e4e6980f17db689d40cebc70e764672f0406e64b46c28ed0b4832d906c1364be\": container with ID starting with e4e6980f17db689d40cebc70e764672f0406e64b46c28ed0b4832d906c1364be not found: ID does not exist" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.657541 4815 scope.go:117] "RemoveContainer" containerID="1906026ad6b21944ed5ec2807d97d4279238deafa5963c99e427e9e0d5ba0459" Feb 25 14:01:05 crc kubenswrapper[4815]: E0225 14:01:05.657932 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1906026ad6b21944ed5ec2807d97d4279238deafa5963c99e427e9e0d5ba0459\": container with ID starting with 1906026ad6b21944ed5ec2807d97d4279238deafa5963c99e427e9e0d5ba0459 not found: ID does not exist" containerID="1906026ad6b21944ed5ec2807d97d4279238deafa5963c99e427e9e0d5ba0459" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.657978 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1906026ad6b21944ed5ec2807d97d4279238deafa5963c99e427e9e0d5ba0459"} err="failed to get container status \"1906026ad6b21944ed5ec2807d97d4279238deafa5963c99e427e9e0d5ba0459\": rpc error: code = NotFound desc = could not find container \"1906026ad6b21944ed5ec2807d97d4279238deafa5963c99e427e9e0d5ba0459\": container with ID starting with 1906026ad6b21944ed5ec2807d97d4279238deafa5963c99e427e9e0d5ba0459 not found: ID does not exist" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.658008 4815 scope.go:117] "RemoveContainer" containerID="da4230c4b4d25b5e6815aaea5ee66ca3f093224f943a39f486ee317460459dab" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.700740 4815 scope.go:117] "RemoveContainer" containerID="7a748c4f9df5ebfec532dd4d918f662e48fdb4322332f811fc839d1db42383b6" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.723369 4815 scope.go:117] "RemoveContainer" containerID="b90f4e2fb5d0ab99a7243fbfffd3eb626f501903334558c6c484b9ec78b25f3a" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.761140 4815 scope.go:117] "RemoveContainer" containerID="da4230c4b4d25b5e6815aaea5ee66ca3f093224f943a39f486ee317460459dab" Feb 25 14:01:05 crc kubenswrapper[4815]: E0225 14:01:05.761962 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da4230c4b4d25b5e6815aaea5ee66ca3f093224f943a39f486ee317460459dab\": container with ID starting with da4230c4b4d25b5e6815aaea5ee66ca3f093224f943a39f486ee317460459dab not found: ID does not exist" containerID="da4230c4b4d25b5e6815aaea5ee66ca3f093224f943a39f486ee317460459dab" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.762003 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da4230c4b4d25b5e6815aaea5ee66ca3f093224f943a39f486ee317460459dab"} err="failed to get container status \"da4230c4b4d25b5e6815aaea5ee66ca3f093224f943a39f486ee317460459dab\": rpc error: code = NotFound desc = could not find container \"da4230c4b4d25b5e6815aaea5ee66ca3f093224f943a39f486ee317460459dab\": container with ID starting with da4230c4b4d25b5e6815aaea5ee66ca3f093224f943a39f486ee317460459dab not found: ID does not exist" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.762037 4815 scope.go:117] "RemoveContainer" containerID="7a748c4f9df5ebfec532dd4d918f662e48fdb4322332f811fc839d1db42383b6" Feb 25 14:01:05 crc kubenswrapper[4815]: E0225 14:01:05.762264 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a748c4f9df5ebfec532dd4d918f662e48fdb4322332f811fc839d1db42383b6\": container with ID starting with 7a748c4f9df5ebfec532dd4d918f662e48fdb4322332f811fc839d1db42383b6 not found: ID does not exist" containerID="7a748c4f9df5ebfec532dd4d918f662e48fdb4322332f811fc839d1db42383b6" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.762299 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a748c4f9df5ebfec532dd4d918f662e48fdb4322332f811fc839d1db42383b6"} err="failed to get container status \"7a748c4f9df5ebfec532dd4d918f662e48fdb4322332f811fc839d1db42383b6\": rpc error: code = NotFound desc = could not find container \"7a748c4f9df5ebfec532dd4d918f662e48fdb4322332f811fc839d1db42383b6\": container with ID starting with 7a748c4f9df5ebfec532dd4d918f662e48fdb4322332f811fc839d1db42383b6 not found: ID does not exist" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.762320 4815 scope.go:117] "RemoveContainer" containerID="b90f4e2fb5d0ab99a7243fbfffd3eb626f501903334558c6c484b9ec78b25f3a" Feb 25 14:01:05 crc kubenswrapper[4815]: E0225 14:01:05.762638 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b90f4e2fb5d0ab99a7243fbfffd3eb626f501903334558c6c484b9ec78b25f3a\": container with ID starting with b90f4e2fb5d0ab99a7243fbfffd3eb626f501903334558c6c484b9ec78b25f3a not found: ID does not exist" containerID="b90f4e2fb5d0ab99a7243fbfffd3eb626f501903334558c6c484b9ec78b25f3a" Feb 25 14:01:05 crc kubenswrapper[4815]: I0225 14:01:05.762700 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b90f4e2fb5d0ab99a7243fbfffd3eb626f501903334558c6c484b9ec78b25f3a"} err="failed to get container status \"b90f4e2fb5d0ab99a7243fbfffd3eb626f501903334558c6c484b9ec78b25f3a\": rpc error: code = NotFound desc = could not find container \"b90f4e2fb5d0ab99a7243fbfffd3eb626f501903334558c6c484b9ec78b25f3a\": container with ID starting with b90f4e2fb5d0ab99a7243fbfffd3eb626f501903334558c6c484b9ec78b25f3a not found: ID does not exist" Feb 25 14:01:06 crc kubenswrapper[4815]: I0225 14:01:06.957958 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12c4d0d5-cdf5-4968-9fde-abe8fac843bd" path="/var/lib/kubelet/pods/12c4d0d5-cdf5-4968-9fde-abe8fac843bd/volumes" Feb 25 14:01:06 crc kubenswrapper[4815]: I0225 14:01:06.959089 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c12242f-0da5-4f59-84e7-77bb094a57dc" path="/var/lib/kubelet/pods/7c12242f-0da5-4f59-84e7-77bb094a57dc/volumes" Feb 25 14:01:10 crc kubenswrapper[4815]: I0225 14:01:10.936203 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 14:01:10 crc kubenswrapper[4815]: E0225 14:01:10.937621 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:01:23 crc kubenswrapper[4815]: I0225 14:01:23.936410 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 14:01:23 crc kubenswrapper[4815]: E0225 14:01:23.938383 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:01:35 crc kubenswrapper[4815]: I0225 14:01:35.936163 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 14:01:35 crc kubenswrapper[4815]: E0225 14:01:35.937493 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:01:48 crc kubenswrapper[4815]: I0225 14:01:48.936464 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 14:01:48 crc kubenswrapper[4815]: E0225 14:01:48.937278 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:01:49 crc kubenswrapper[4815]: I0225 14:01:49.009795 4815 generic.go:334] "Generic (PLEG): container finished" podID="1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6" containerID="afc4247d75b19ec77a9d3cb63c1fc28ca373ab90e03b0266b73d98ab4d113a48" exitCode=0 Feb 25 14:01:49 crc kubenswrapper[4815]: I0225 14:01:49.009876 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" event={"ID":"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6","Type":"ContainerDied","Data":"afc4247d75b19ec77a9d3cb63c1fc28ca373ab90e03b0266b73d98ab4d113a48"} Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.445653 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.633614 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-telemetry-combined-ca-bundle\") pod \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.633692 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-inventory\") pod \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.633742 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-1\") pod \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.633807 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ssh-key-openstack-edpm-ipam\") pod \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.633844 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkfdw\" (UniqueName: \"kubernetes.io/projected/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-kube-api-access-dkfdw\") pod \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.633931 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-2\") pod \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.633993 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-0\") pod \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\" (UID: \"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6\") " Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.642700 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-kube-api-access-dkfdw" (OuterVolumeSpecName: "kube-api-access-dkfdw") pod "1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6" (UID: "1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6"). InnerVolumeSpecName "kube-api-access-dkfdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.646082 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6" (UID: "1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.666616 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-inventory" (OuterVolumeSpecName: "inventory") pod "1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6" (UID: "1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.674624 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6" (UID: "1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.675700 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6" (UID: "1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.691149 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6" (UID: "1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.695112 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6" (UID: "1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.736608 4815 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.736671 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.736692 4815 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.736711 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.736734 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkfdw\" (UniqueName: \"kubernetes.io/projected/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-kube-api-access-dkfdw\") on node \"crc\" DevicePath \"\"" Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.736751 4815 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Feb 25 14:01:50 crc kubenswrapper[4815]: I0225 14:01:50.736773 4815 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Feb 25 14:01:51 crc kubenswrapper[4815]: I0225 14:01:51.036308 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" event={"ID":"1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6","Type":"ContainerDied","Data":"d3a1f5d845196ee333027d9c993ad2211817eb1109c7ec059a527f87fb209d20"} Feb 25 14:01:51 crc kubenswrapper[4815]: I0225 14:01:51.036373 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3a1f5d845196ee333027d9c993ad2211817eb1109c7ec059a527f87fb209d20" Feb 25 14:01:51 crc kubenswrapper[4815]: I0225 14:01:51.036430 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-47nxh" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.148358 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533802-mkd6f"] Feb 25 14:02:00 crc kubenswrapper[4815]: E0225 14:02:00.149594 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c4d0d5-cdf5-4968-9fde-abe8fac843bd" containerName="registry-server" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.149616 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c4d0d5-cdf5-4968-9fde-abe8fac843bd" containerName="registry-server" Feb 25 14:02:00 crc kubenswrapper[4815]: E0225 14:02:00.149644 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c12242f-0da5-4f59-84e7-77bb094a57dc" containerName="registry-server" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.149655 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c12242f-0da5-4f59-84e7-77bb094a57dc" containerName="registry-server" Feb 25 14:02:00 crc kubenswrapper[4815]: E0225 14:02:00.149685 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c12242f-0da5-4f59-84e7-77bb094a57dc" containerName="extract-content" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.149695 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c12242f-0da5-4f59-84e7-77bb094a57dc" containerName="extract-content" Feb 25 14:02:00 crc kubenswrapper[4815]: E0225 14:02:00.149721 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c4d0d5-cdf5-4968-9fde-abe8fac843bd" containerName="extract-utilities" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.149731 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c4d0d5-cdf5-4968-9fde-abe8fac843bd" containerName="extract-utilities" Feb 25 14:02:00 crc kubenswrapper[4815]: E0225 14:02:00.149751 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99ed322b-3f0d-4adc-8244-07af594b14e7" containerName="keystone-cron" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.149761 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="99ed322b-3f0d-4adc-8244-07af594b14e7" containerName="keystone-cron" Feb 25 14:02:00 crc kubenswrapper[4815]: E0225 14:02:00.149786 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.149798 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 25 14:02:00 crc kubenswrapper[4815]: E0225 14:02:00.149813 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c12242f-0da5-4f59-84e7-77bb094a57dc" containerName="extract-utilities" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.149824 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c12242f-0da5-4f59-84e7-77bb094a57dc" containerName="extract-utilities" Feb 25 14:02:00 crc kubenswrapper[4815]: E0225 14:02:00.149847 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c4d0d5-cdf5-4968-9fde-abe8fac843bd" containerName="extract-content" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.149857 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c4d0d5-cdf5-4968-9fde-abe8fac843bd" containerName="extract-content" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.150147 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.150172 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="12c4d0d5-cdf5-4968-9fde-abe8fac843bd" containerName="registry-server" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.150187 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c12242f-0da5-4f59-84e7-77bb094a57dc" containerName="registry-server" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.150229 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="99ed322b-3f0d-4adc-8244-07af594b14e7" containerName="keystone-cron" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.151281 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533802-mkd6f" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.154087 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.154164 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.154259 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.160944 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533802-mkd6f"] Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.241974 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxm4h\" (UniqueName: \"kubernetes.io/projected/e236e592-fc11-48e8-a1d4-65178884f48a-kube-api-access-pxm4h\") pod \"auto-csr-approver-29533802-mkd6f\" (UID: \"e236e592-fc11-48e8-a1d4-65178884f48a\") " pod="openshift-infra/auto-csr-approver-29533802-mkd6f" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.343573 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxm4h\" (UniqueName: \"kubernetes.io/projected/e236e592-fc11-48e8-a1d4-65178884f48a-kube-api-access-pxm4h\") pod \"auto-csr-approver-29533802-mkd6f\" (UID: \"e236e592-fc11-48e8-a1d4-65178884f48a\") " pod="openshift-infra/auto-csr-approver-29533802-mkd6f" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.361269 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxm4h\" (UniqueName: \"kubernetes.io/projected/e236e592-fc11-48e8-a1d4-65178884f48a-kube-api-access-pxm4h\") pod \"auto-csr-approver-29533802-mkd6f\" (UID: \"e236e592-fc11-48e8-a1d4-65178884f48a\") " pod="openshift-infra/auto-csr-approver-29533802-mkd6f" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.476754 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533802-mkd6f" Feb 25 14:02:00 crc kubenswrapper[4815]: I0225 14:02:00.979120 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533802-mkd6f"] Feb 25 14:02:01 crc kubenswrapper[4815]: I0225 14:02:01.151813 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533802-mkd6f" event={"ID":"e236e592-fc11-48e8-a1d4-65178884f48a","Type":"ContainerStarted","Data":"a12d7755fbb265bb4a0b35b221fa68c16d5bf038a9ead679da3fc7a8ac511bf4"} Feb 25 14:02:03 crc kubenswrapper[4815]: I0225 14:02:03.181749 4815 generic.go:334] "Generic (PLEG): container finished" podID="e236e592-fc11-48e8-a1d4-65178884f48a" containerID="111efb696e991623cb96fb6d345e3d2aeb0175e78ffe239cfddec2911985a3d2" exitCode=0 Feb 25 14:02:03 crc kubenswrapper[4815]: I0225 14:02:03.181843 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533802-mkd6f" event={"ID":"e236e592-fc11-48e8-a1d4-65178884f48a","Type":"ContainerDied","Data":"111efb696e991623cb96fb6d345e3d2aeb0175e78ffe239cfddec2911985a3d2"} Feb 25 14:02:03 crc kubenswrapper[4815]: I0225 14:02:03.942798 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 14:02:03 crc kubenswrapper[4815]: E0225 14:02:03.943846 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:02:04 crc kubenswrapper[4815]: I0225 14:02:04.557166 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533802-mkd6f" Feb 25 14:02:04 crc kubenswrapper[4815]: I0225 14:02:04.727845 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxm4h\" (UniqueName: \"kubernetes.io/projected/e236e592-fc11-48e8-a1d4-65178884f48a-kube-api-access-pxm4h\") pod \"e236e592-fc11-48e8-a1d4-65178884f48a\" (UID: \"e236e592-fc11-48e8-a1d4-65178884f48a\") " Feb 25 14:02:04 crc kubenswrapper[4815]: I0225 14:02:04.734363 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e236e592-fc11-48e8-a1d4-65178884f48a-kube-api-access-pxm4h" (OuterVolumeSpecName: "kube-api-access-pxm4h") pod "e236e592-fc11-48e8-a1d4-65178884f48a" (UID: "e236e592-fc11-48e8-a1d4-65178884f48a"). InnerVolumeSpecName "kube-api-access-pxm4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:02:04 crc kubenswrapper[4815]: I0225 14:02:04.829654 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxm4h\" (UniqueName: \"kubernetes.io/projected/e236e592-fc11-48e8-a1d4-65178884f48a-kube-api-access-pxm4h\") on node \"crc\" DevicePath \"\"" Feb 25 14:02:05 crc kubenswrapper[4815]: I0225 14:02:05.205405 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533802-mkd6f" event={"ID":"e236e592-fc11-48e8-a1d4-65178884f48a","Type":"ContainerDied","Data":"a12d7755fbb265bb4a0b35b221fa68c16d5bf038a9ead679da3fc7a8ac511bf4"} Feb 25 14:02:05 crc kubenswrapper[4815]: I0225 14:02:05.205490 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a12d7755fbb265bb4a0b35b221fa68c16d5bf038a9ead679da3fc7a8ac511bf4" Feb 25 14:02:05 crc kubenswrapper[4815]: I0225 14:02:05.205533 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533802-mkd6f" Feb 25 14:02:05 crc kubenswrapper[4815]: I0225 14:02:05.654491 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533796-68pz9"] Feb 25 14:02:05 crc kubenswrapper[4815]: I0225 14:02:05.664828 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533796-68pz9"] Feb 25 14:02:06 crc kubenswrapper[4815]: I0225 14:02:06.959425 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d5d295f-9999-4500-af75-c46724b26d89" path="/var/lib/kubelet/pods/7d5d295f-9999-4500-af75-c46724b26d89/volumes" Feb 25 14:02:14 crc kubenswrapper[4815]: I0225 14:02:14.936379 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 14:02:14 crc kubenswrapper[4815]: E0225 14:02:14.937279 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:02:27 crc kubenswrapper[4815]: I0225 14:02:27.936360 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 14:02:27 crc kubenswrapper[4815]: E0225 14:02:27.937724 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:02:41 crc kubenswrapper[4815]: I0225 14:02:41.936356 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 14:02:41 crc kubenswrapper[4815]: E0225 14:02:41.937060 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.219661 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Feb 25 14:02:47 crc kubenswrapper[4815]: E0225 14:02:47.220809 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e236e592-fc11-48e8-a1d4-65178884f48a" containerName="oc" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.220827 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e236e592-fc11-48e8-a1d4-65178884f48a" containerName="oc" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.221051 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e236e592-fc11-48e8-a1d4-65178884f48a" containerName="oc" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.221822 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.226329 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.226728 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.226795 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-sqqf7" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.226912 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.239184 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.328160 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.328226 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.328283 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqkvk\" (UniqueName: \"kubernetes.io/projected/d23e8f04-361c-4227-a5c9-bfcd4801f771-kube-api-access-lqkvk\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.328552 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d23e8f04-361c-4227-a5c9-bfcd4801f771-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.328677 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d23e8f04-361c-4227-a5c9-bfcd4801f771-config-data\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.328720 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d23e8f04-361c-4227-a5c9-bfcd4801f771-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.328738 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.328938 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.329031 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d23e8f04-361c-4227-a5c9-bfcd4801f771-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.430670 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d23e8f04-361c-4227-a5c9-bfcd4801f771-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.430761 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.430859 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.430923 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d23e8f04-361c-4227-a5c9-bfcd4801f771-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.431024 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.431112 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.431605 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.431996 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d23e8f04-361c-4227-a5c9-bfcd4801f771-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.432095 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqkvk\" (UniqueName: \"kubernetes.io/projected/d23e8f04-361c-4227-a5c9-bfcd4801f771-kube-api-access-lqkvk\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.432302 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d23e8f04-361c-4227-a5c9-bfcd4801f771-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.432703 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d23e8f04-361c-4227-a5c9-bfcd4801f771-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.433281 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d23e8f04-361c-4227-a5c9-bfcd4801f771-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.433635 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d23e8f04-361c-4227-a5c9-bfcd4801f771-config-data\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.435131 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d23e8f04-361c-4227-a5c9-bfcd4801f771-config-data\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.438223 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.443372 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.444215 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.451040 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqkvk\" (UniqueName: \"kubernetes.io/projected/d23e8f04-361c-4227-a5c9-bfcd4801f771-kube-api-access-lqkvk\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.460558 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.563900 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 25 14:02:47 crc kubenswrapper[4815]: I0225 14:02:47.874821 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 25 14:02:48 crc kubenswrapper[4815]: I0225 14:02:48.646482 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d23e8f04-361c-4227-a5c9-bfcd4801f771","Type":"ContainerStarted","Data":"84b2df7e536fa12971c21c2080bccea901c67f26be53d99e35ec961c4e9d9a4b"} Feb 25 14:02:48 crc kubenswrapper[4815]: I0225 14:02:48.844755 4815 scope.go:117] "RemoveContainer" containerID="7b9bbcacfb5d26bb4c9bad6ee1cd8df0a2010a679261982c3a3c7ddd27bdb0bd" Feb 25 14:02:56 crc kubenswrapper[4815]: I0225 14:02:56.935954 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 14:02:58 crc kubenswrapper[4815]: I0225 14:02:58.757901 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerStarted","Data":"aa536f69b8a87db0cb31810395d99769259dc48b08c1c927d09e2d481a2a47df"} Feb 25 14:03:14 crc kubenswrapper[4815]: E0225 14:03:14.094693 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Feb 25 14:03:14 crc kubenswrapper[4815]: E0225 14:03:14.095645 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lqkvk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(d23e8f04-361c-4227-a5c9-bfcd4801f771): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 14:03:14 crc kubenswrapper[4815]: E0225 14:03:14.096837 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="d23e8f04-361c-4227-a5c9-bfcd4801f771" Feb 25 14:03:14 crc kubenswrapper[4815]: E0225 14:03:14.953474 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="d23e8f04-361c-4227-a5c9-bfcd4801f771" Feb 25 14:03:32 crc kubenswrapper[4815]: I0225 14:03:32.154543 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d23e8f04-361c-4227-a5c9-bfcd4801f771","Type":"ContainerStarted","Data":"3cf1f2d98cbdd054697aa239331c4744120fbfbd01368da1d5dbd4c6ce8bd28f"} Feb 25 14:03:32 crc kubenswrapper[4815]: I0225 14:03:32.183782 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.624966972 podStartE2EDuration="46.18376508s" podCreationTimestamp="2026-02-25 14:02:46 +0000 UTC" firstStartedPulling="2026-02-25 14:02:47.880335109 +0000 UTC m=+2525.681433163" lastFinishedPulling="2026-02-25 14:03:30.439133177 +0000 UTC m=+2568.240231271" observedRunningTime="2026-02-25 14:03:32.181252842 +0000 UTC m=+2569.982350906" watchObservedRunningTime="2026-02-25 14:03:32.18376508 +0000 UTC m=+2569.984863134" Feb 25 14:04:00 crc kubenswrapper[4815]: I0225 14:04:00.164784 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533804-rpvkq"] Feb 25 14:04:00 crc kubenswrapper[4815]: I0225 14:04:00.167006 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533804-rpvkq" Feb 25 14:04:00 crc kubenswrapper[4815]: I0225 14:04:00.170131 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 14:04:00 crc kubenswrapper[4815]: I0225 14:04:00.170182 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 14:04:00 crc kubenswrapper[4815]: I0225 14:04:00.170887 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 14:04:00 crc kubenswrapper[4815]: I0225 14:04:00.179222 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533804-rpvkq"] Feb 25 14:04:00 crc kubenswrapper[4815]: I0225 14:04:00.282813 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntt5b\" (UniqueName: \"kubernetes.io/projected/8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d-kube-api-access-ntt5b\") pod \"auto-csr-approver-29533804-rpvkq\" (UID: \"8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d\") " pod="openshift-infra/auto-csr-approver-29533804-rpvkq" Feb 25 14:04:00 crc kubenswrapper[4815]: I0225 14:04:00.385094 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntt5b\" (UniqueName: \"kubernetes.io/projected/8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d-kube-api-access-ntt5b\") pod \"auto-csr-approver-29533804-rpvkq\" (UID: \"8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d\") " pod="openshift-infra/auto-csr-approver-29533804-rpvkq" Feb 25 14:04:00 crc kubenswrapper[4815]: I0225 14:04:00.418082 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntt5b\" (UniqueName: \"kubernetes.io/projected/8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d-kube-api-access-ntt5b\") pod \"auto-csr-approver-29533804-rpvkq\" (UID: \"8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d\") " pod="openshift-infra/auto-csr-approver-29533804-rpvkq" Feb 25 14:04:00 crc kubenswrapper[4815]: I0225 14:04:00.486073 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533804-rpvkq" Feb 25 14:04:01 crc kubenswrapper[4815]: I0225 14:04:01.142689 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533804-rpvkq"] Feb 25 14:04:01 crc kubenswrapper[4815]: W0225 14:04:01.145184 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fa0c6ba_5bb6_4fcd_88d0_3c572a4c761d.slice/crio-52113e83bea467aeb89ed37d1a1aa312a82b0adb8ca6bd712289cc01969dfa6b WatchSource:0}: Error finding container 52113e83bea467aeb89ed37d1a1aa312a82b0adb8ca6bd712289cc01969dfa6b: Status 404 returned error can't find the container with id 52113e83bea467aeb89ed37d1a1aa312a82b0adb8ca6bd712289cc01969dfa6b Feb 25 14:04:01 crc kubenswrapper[4815]: I0225 14:04:01.500983 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533804-rpvkq" event={"ID":"8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d","Type":"ContainerStarted","Data":"52113e83bea467aeb89ed37d1a1aa312a82b0adb8ca6bd712289cc01969dfa6b"} Feb 25 14:04:02 crc kubenswrapper[4815]: I0225 14:04:02.511120 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533804-rpvkq" event={"ID":"8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d","Type":"ContainerStarted","Data":"f64b6650bc4aaa76893d99c1eefaa798e69267d5d795d21dbfed923c703e4c9a"} Feb 25 14:04:02 crc kubenswrapper[4815]: I0225 14:04:02.532462 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533804-rpvkq" podStartSLOduration=1.704434104 podStartE2EDuration="2.532441118s" podCreationTimestamp="2026-02-25 14:04:00 +0000 UTC" firstStartedPulling="2026-02-25 14:04:01.148403692 +0000 UTC m=+2598.949501746" lastFinishedPulling="2026-02-25 14:04:01.976410696 +0000 UTC m=+2599.777508760" observedRunningTime="2026-02-25 14:04:02.525316086 +0000 UTC m=+2600.326414150" watchObservedRunningTime="2026-02-25 14:04:02.532441118 +0000 UTC m=+2600.333539172" Feb 25 14:04:03 crc kubenswrapper[4815]: I0225 14:04:03.522608 4815 generic.go:334] "Generic (PLEG): container finished" podID="8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d" containerID="f64b6650bc4aaa76893d99c1eefaa798e69267d5d795d21dbfed923c703e4c9a" exitCode=0 Feb 25 14:04:03 crc kubenswrapper[4815]: I0225 14:04:03.522651 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533804-rpvkq" event={"ID":"8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d","Type":"ContainerDied","Data":"f64b6650bc4aaa76893d99c1eefaa798e69267d5d795d21dbfed923c703e4c9a"} Feb 25 14:04:04 crc kubenswrapper[4815]: I0225 14:04:04.924074 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533804-rpvkq" Feb 25 14:04:05 crc kubenswrapper[4815]: I0225 14:04:05.078074 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntt5b\" (UniqueName: \"kubernetes.io/projected/8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d-kube-api-access-ntt5b\") pod \"8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d\" (UID: \"8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d\") " Feb 25 14:04:05 crc kubenswrapper[4815]: I0225 14:04:05.083795 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d-kube-api-access-ntt5b" (OuterVolumeSpecName: "kube-api-access-ntt5b") pod "8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d" (UID: "8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d"). InnerVolumeSpecName "kube-api-access-ntt5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:04:05 crc kubenswrapper[4815]: I0225 14:04:05.180659 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntt5b\" (UniqueName: \"kubernetes.io/projected/8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d-kube-api-access-ntt5b\") on node \"crc\" DevicePath \"\"" Feb 25 14:04:05 crc kubenswrapper[4815]: I0225 14:04:05.546959 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533804-rpvkq" event={"ID":"8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d","Type":"ContainerDied","Data":"52113e83bea467aeb89ed37d1a1aa312a82b0adb8ca6bd712289cc01969dfa6b"} Feb 25 14:04:05 crc kubenswrapper[4815]: I0225 14:04:05.547003 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52113e83bea467aeb89ed37d1a1aa312a82b0adb8ca6bd712289cc01969dfa6b" Feb 25 14:04:05 crc kubenswrapper[4815]: I0225 14:04:05.547063 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533804-rpvkq" Feb 25 14:04:05 crc kubenswrapper[4815]: I0225 14:04:05.608794 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533798-7bnr8"] Feb 25 14:04:05 crc kubenswrapper[4815]: I0225 14:04:05.619947 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533798-7bnr8"] Feb 25 14:04:06 crc kubenswrapper[4815]: I0225 14:04:06.954619 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e7253d5-5810-4271-a58b-040a3eb09d82" path="/var/lib/kubelet/pods/7e7253d5-5810-4271-a58b-040a3eb09d82/volumes" Feb 25 14:04:48 crc kubenswrapper[4815]: I0225 14:04:48.952941 4815 scope.go:117] "RemoveContainer" containerID="c94113bf1e90f8bff4a9480c92916ba73c04497222227dba629dc1ea749df28b" Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.074055 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4745t"] Feb 25 14:05:21 crc kubenswrapper[4815]: E0225 14:05:21.075083 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d" containerName="oc" Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.075101 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d" containerName="oc" Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.075344 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d" containerName="oc" Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.077064 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4745t" Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.096113 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4745t"] Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.184222 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe21655-da41-4139-ad53-0c6f04a3477e-utilities\") pod \"redhat-marketplace-4745t\" (UID: \"9fe21655-da41-4139-ad53-0c6f04a3477e\") " pod="openshift-marketplace/redhat-marketplace-4745t" Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.184296 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4shvj\" (UniqueName: \"kubernetes.io/projected/9fe21655-da41-4139-ad53-0c6f04a3477e-kube-api-access-4shvj\") pod \"redhat-marketplace-4745t\" (UID: \"9fe21655-da41-4139-ad53-0c6f04a3477e\") " pod="openshift-marketplace/redhat-marketplace-4745t" Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.184759 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe21655-da41-4139-ad53-0c6f04a3477e-catalog-content\") pod \"redhat-marketplace-4745t\" (UID: \"9fe21655-da41-4139-ad53-0c6f04a3477e\") " pod="openshift-marketplace/redhat-marketplace-4745t" Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.286741 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe21655-da41-4139-ad53-0c6f04a3477e-catalog-content\") pod \"redhat-marketplace-4745t\" (UID: \"9fe21655-da41-4139-ad53-0c6f04a3477e\") " pod="openshift-marketplace/redhat-marketplace-4745t" Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.286884 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe21655-da41-4139-ad53-0c6f04a3477e-utilities\") pod \"redhat-marketplace-4745t\" (UID: \"9fe21655-da41-4139-ad53-0c6f04a3477e\") " pod="openshift-marketplace/redhat-marketplace-4745t" Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.286938 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4shvj\" (UniqueName: \"kubernetes.io/projected/9fe21655-da41-4139-ad53-0c6f04a3477e-kube-api-access-4shvj\") pod \"redhat-marketplace-4745t\" (UID: \"9fe21655-da41-4139-ad53-0c6f04a3477e\") " pod="openshift-marketplace/redhat-marketplace-4745t" Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.287693 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe21655-da41-4139-ad53-0c6f04a3477e-catalog-content\") pod \"redhat-marketplace-4745t\" (UID: \"9fe21655-da41-4139-ad53-0c6f04a3477e\") " pod="openshift-marketplace/redhat-marketplace-4745t" Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.287845 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe21655-da41-4139-ad53-0c6f04a3477e-utilities\") pod \"redhat-marketplace-4745t\" (UID: \"9fe21655-da41-4139-ad53-0c6f04a3477e\") " pod="openshift-marketplace/redhat-marketplace-4745t" Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.312902 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4shvj\" (UniqueName: \"kubernetes.io/projected/9fe21655-da41-4139-ad53-0c6f04a3477e-kube-api-access-4shvj\") pod \"redhat-marketplace-4745t\" (UID: \"9fe21655-da41-4139-ad53-0c6f04a3477e\") " pod="openshift-marketplace/redhat-marketplace-4745t" Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.341676 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.341731 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.411489 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4745t" Feb 25 14:05:21 crc kubenswrapper[4815]: I0225 14:05:21.923301 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4745t"] Feb 25 14:05:21 crc kubenswrapper[4815]: W0225 14:05:21.924044 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fe21655_da41_4139_ad53_0c6f04a3477e.slice/crio-0963ea60332d0a760ceb661c0d94080e365a48a1b6adc8f607b12683dc4b25da WatchSource:0}: Error finding container 0963ea60332d0a760ceb661c0d94080e365a48a1b6adc8f607b12683dc4b25da: Status 404 returned error can't find the container with id 0963ea60332d0a760ceb661c0d94080e365a48a1b6adc8f607b12683dc4b25da Feb 25 14:05:22 crc kubenswrapper[4815]: I0225 14:05:22.261081 4815 generic.go:334] "Generic (PLEG): container finished" podID="9fe21655-da41-4139-ad53-0c6f04a3477e" containerID="7d96cf11ce8fa9093be8c15317217c1e1949481bb3a6f2658ccfa848d0f2e9fd" exitCode=0 Feb 25 14:05:22 crc kubenswrapper[4815]: I0225 14:05:22.261149 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4745t" event={"ID":"9fe21655-da41-4139-ad53-0c6f04a3477e","Type":"ContainerDied","Data":"7d96cf11ce8fa9093be8c15317217c1e1949481bb3a6f2658ccfa848d0f2e9fd"} Feb 25 14:05:22 crc kubenswrapper[4815]: I0225 14:05:22.261367 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4745t" event={"ID":"9fe21655-da41-4139-ad53-0c6f04a3477e","Type":"ContainerStarted","Data":"0963ea60332d0a760ceb661c0d94080e365a48a1b6adc8f607b12683dc4b25da"} Feb 25 14:05:23 crc kubenswrapper[4815]: I0225 14:05:23.272495 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4745t" event={"ID":"9fe21655-da41-4139-ad53-0c6f04a3477e","Type":"ContainerStarted","Data":"39262444671f61191f42e2bd20ac99c0b65f11180a606ad40d85cd72a3bc4ba2"} Feb 25 14:05:24 crc kubenswrapper[4815]: I0225 14:05:24.285231 4815 generic.go:334] "Generic (PLEG): container finished" podID="9fe21655-da41-4139-ad53-0c6f04a3477e" containerID="39262444671f61191f42e2bd20ac99c0b65f11180a606ad40d85cd72a3bc4ba2" exitCode=0 Feb 25 14:05:24 crc kubenswrapper[4815]: I0225 14:05:24.285358 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4745t" event={"ID":"9fe21655-da41-4139-ad53-0c6f04a3477e","Type":"ContainerDied","Data":"39262444671f61191f42e2bd20ac99c0b65f11180a606ad40d85cd72a3bc4ba2"} Feb 25 14:05:24 crc kubenswrapper[4815]: I0225 14:05:24.285625 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4745t" event={"ID":"9fe21655-da41-4139-ad53-0c6f04a3477e","Type":"ContainerStarted","Data":"be3134d4534ac762b3da7aef8aac0e43dd701f9a333236c015ba7ed5d496814a"} Feb 25 14:05:24 crc kubenswrapper[4815]: I0225 14:05:24.306999 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4745t" podStartSLOduration=1.906122785 podStartE2EDuration="3.306972479s" podCreationTimestamp="2026-02-25 14:05:21 +0000 UTC" firstStartedPulling="2026-02-25 14:05:22.264604167 +0000 UTC m=+2680.065702261" lastFinishedPulling="2026-02-25 14:05:23.665453901 +0000 UTC m=+2681.466551955" observedRunningTime="2026-02-25 14:05:24.302495708 +0000 UTC m=+2682.103593802" watchObservedRunningTime="2026-02-25 14:05:24.306972479 +0000 UTC m=+2682.108070553" Feb 25 14:05:31 crc kubenswrapper[4815]: I0225 14:05:31.412646 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4745t" Feb 25 14:05:31 crc kubenswrapper[4815]: I0225 14:05:31.413224 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4745t" Feb 25 14:05:31 crc kubenswrapper[4815]: I0225 14:05:31.465916 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4745t" Feb 25 14:05:32 crc kubenswrapper[4815]: I0225 14:05:32.438487 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4745t" Feb 25 14:05:32 crc kubenswrapper[4815]: I0225 14:05:32.489488 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4745t"] Feb 25 14:05:34 crc kubenswrapper[4815]: I0225 14:05:34.385189 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4745t" podUID="9fe21655-da41-4139-ad53-0c6f04a3477e" containerName="registry-server" containerID="cri-o://be3134d4534ac762b3da7aef8aac0e43dd701f9a333236c015ba7ed5d496814a" gracePeriod=2 Feb 25 14:05:34 crc kubenswrapper[4815]: I0225 14:05:34.962443 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4745t" Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.090398 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe21655-da41-4139-ad53-0c6f04a3477e-catalog-content\") pod \"9fe21655-da41-4139-ad53-0c6f04a3477e\" (UID: \"9fe21655-da41-4139-ad53-0c6f04a3477e\") " Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.090489 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe21655-da41-4139-ad53-0c6f04a3477e-utilities\") pod \"9fe21655-da41-4139-ad53-0c6f04a3477e\" (UID: \"9fe21655-da41-4139-ad53-0c6f04a3477e\") " Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.090581 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4shvj\" (UniqueName: \"kubernetes.io/projected/9fe21655-da41-4139-ad53-0c6f04a3477e-kube-api-access-4shvj\") pod \"9fe21655-da41-4139-ad53-0c6f04a3477e\" (UID: \"9fe21655-da41-4139-ad53-0c6f04a3477e\") " Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.091616 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fe21655-da41-4139-ad53-0c6f04a3477e-utilities" (OuterVolumeSpecName: "utilities") pod "9fe21655-da41-4139-ad53-0c6f04a3477e" (UID: "9fe21655-da41-4139-ad53-0c6f04a3477e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.096075 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fe21655-da41-4139-ad53-0c6f04a3477e-kube-api-access-4shvj" (OuterVolumeSpecName: "kube-api-access-4shvj") pod "9fe21655-da41-4139-ad53-0c6f04a3477e" (UID: "9fe21655-da41-4139-ad53-0c6f04a3477e"). InnerVolumeSpecName "kube-api-access-4shvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.113328 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fe21655-da41-4139-ad53-0c6f04a3477e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9fe21655-da41-4139-ad53-0c6f04a3477e" (UID: "9fe21655-da41-4139-ad53-0c6f04a3477e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.193196 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe21655-da41-4139-ad53-0c6f04a3477e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.193227 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe21655-da41-4139-ad53-0c6f04a3477e-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.193236 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4shvj\" (UniqueName: \"kubernetes.io/projected/9fe21655-da41-4139-ad53-0c6f04a3477e-kube-api-access-4shvj\") on node \"crc\" DevicePath \"\"" Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.400602 4815 generic.go:334] "Generic (PLEG): container finished" podID="9fe21655-da41-4139-ad53-0c6f04a3477e" containerID="be3134d4534ac762b3da7aef8aac0e43dd701f9a333236c015ba7ed5d496814a" exitCode=0 Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.400643 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4745t" event={"ID":"9fe21655-da41-4139-ad53-0c6f04a3477e","Type":"ContainerDied","Data":"be3134d4534ac762b3da7aef8aac0e43dd701f9a333236c015ba7ed5d496814a"} Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.400669 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4745t" event={"ID":"9fe21655-da41-4139-ad53-0c6f04a3477e","Type":"ContainerDied","Data":"0963ea60332d0a760ceb661c0d94080e365a48a1b6adc8f607b12683dc4b25da"} Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.400684 4815 scope.go:117] "RemoveContainer" containerID="be3134d4534ac762b3da7aef8aac0e43dd701f9a333236c015ba7ed5d496814a" Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.400788 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4745t" Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.439872 4815 scope.go:117] "RemoveContainer" containerID="39262444671f61191f42e2bd20ac99c0b65f11180a606ad40d85cd72a3bc4ba2" Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.467260 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4745t"] Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.469736 4815 scope.go:117] "RemoveContainer" containerID="7d96cf11ce8fa9093be8c15317217c1e1949481bb3a6f2658ccfa848d0f2e9fd" Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.487213 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4745t"] Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.547071 4815 scope.go:117] "RemoveContainer" containerID="be3134d4534ac762b3da7aef8aac0e43dd701f9a333236c015ba7ed5d496814a" Feb 25 14:05:35 crc kubenswrapper[4815]: E0225 14:05:35.547542 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be3134d4534ac762b3da7aef8aac0e43dd701f9a333236c015ba7ed5d496814a\": container with ID starting with be3134d4534ac762b3da7aef8aac0e43dd701f9a333236c015ba7ed5d496814a not found: ID does not exist" containerID="be3134d4534ac762b3da7aef8aac0e43dd701f9a333236c015ba7ed5d496814a" Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.547582 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be3134d4534ac762b3da7aef8aac0e43dd701f9a333236c015ba7ed5d496814a"} err="failed to get container status \"be3134d4534ac762b3da7aef8aac0e43dd701f9a333236c015ba7ed5d496814a\": rpc error: code = NotFound desc = could not find container \"be3134d4534ac762b3da7aef8aac0e43dd701f9a333236c015ba7ed5d496814a\": container with ID starting with be3134d4534ac762b3da7aef8aac0e43dd701f9a333236c015ba7ed5d496814a not found: ID does not exist" Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.547607 4815 scope.go:117] "RemoveContainer" containerID="39262444671f61191f42e2bd20ac99c0b65f11180a606ad40d85cd72a3bc4ba2" Feb 25 14:05:35 crc kubenswrapper[4815]: E0225 14:05:35.551010 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39262444671f61191f42e2bd20ac99c0b65f11180a606ad40d85cd72a3bc4ba2\": container with ID starting with 39262444671f61191f42e2bd20ac99c0b65f11180a606ad40d85cd72a3bc4ba2 not found: ID does not exist" containerID="39262444671f61191f42e2bd20ac99c0b65f11180a606ad40d85cd72a3bc4ba2" Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.551180 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39262444671f61191f42e2bd20ac99c0b65f11180a606ad40d85cd72a3bc4ba2"} err="failed to get container status \"39262444671f61191f42e2bd20ac99c0b65f11180a606ad40d85cd72a3bc4ba2\": rpc error: code = NotFound desc = could not find container \"39262444671f61191f42e2bd20ac99c0b65f11180a606ad40d85cd72a3bc4ba2\": container with ID starting with 39262444671f61191f42e2bd20ac99c0b65f11180a606ad40d85cd72a3bc4ba2 not found: ID does not exist" Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.551294 4815 scope.go:117] "RemoveContainer" containerID="7d96cf11ce8fa9093be8c15317217c1e1949481bb3a6f2658ccfa848d0f2e9fd" Feb 25 14:05:35 crc kubenswrapper[4815]: E0225 14:05:35.551618 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d96cf11ce8fa9093be8c15317217c1e1949481bb3a6f2658ccfa848d0f2e9fd\": container with ID starting with 7d96cf11ce8fa9093be8c15317217c1e1949481bb3a6f2658ccfa848d0f2e9fd not found: ID does not exist" containerID="7d96cf11ce8fa9093be8c15317217c1e1949481bb3a6f2658ccfa848d0f2e9fd" Feb 25 14:05:35 crc kubenswrapper[4815]: I0225 14:05:35.551682 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d96cf11ce8fa9093be8c15317217c1e1949481bb3a6f2658ccfa848d0f2e9fd"} err="failed to get container status \"7d96cf11ce8fa9093be8c15317217c1e1949481bb3a6f2658ccfa848d0f2e9fd\": rpc error: code = NotFound desc = could not find container \"7d96cf11ce8fa9093be8c15317217c1e1949481bb3a6f2658ccfa848d0f2e9fd\": container with ID starting with 7d96cf11ce8fa9093be8c15317217c1e1949481bb3a6f2658ccfa848d0f2e9fd not found: ID does not exist" Feb 25 14:05:36 crc kubenswrapper[4815]: I0225 14:05:36.948514 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fe21655-da41-4139-ad53-0c6f04a3477e" path="/var/lib/kubelet/pods/9fe21655-da41-4139-ad53-0c6f04a3477e/volumes" Feb 25 14:05:51 crc kubenswrapper[4815]: I0225 14:05:51.341987 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 14:05:51 crc kubenswrapper[4815]: I0225 14:05:51.342623 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 14:06:00 crc kubenswrapper[4815]: I0225 14:06:00.163045 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533806-l2585"] Feb 25 14:06:00 crc kubenswrapper[4815]: E0225 14:06:00.163961 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fe21655-da41-4139-ad53-0c6f04a3477e" containerName="extract-content" Feb 25 14:06:00 crc kubenswrapper[4815]: I0225 14:06:00.163976 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fe21655-da41-4139-ad53-0c6f04a3477e" containerName="extract-content" Feb 25 14:06:00 crc kubenswrapper[4815]: E0225 14:06:00.163985 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fe21655-da41-4139-ad53-0c6f04a3477e" containerName="registry-server" Feb 25 14:06:00 crc kubenswrapper[4815]: I0225 14:06:00.163992 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fe21655-da41-4139-ad53-0c6f04a3477e" containerName="registry-server" Feb 25 14:06:00 crc kubenswrapper[4815]: E0225 14:06:00.164033 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fe21655-da41-4139-ad53-0c6f04a3477e" containerName="extract-utilities" Feb 25 14:06:00 crc kubenswrapper[4815]: I0225 14:06:00.164042 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fe21655-da41-4139-ad53-0c6f04a3477e" containerName="extract-utilities" Feb 25 14:06:00 crc kubenswrapper[4815]: I0225 14:06:00.164251 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fe21655-da41-4139-ad53-0c6f04a3477e" containerName="registry-server" Feb 25 14:06:00 crc kubenswrapper[4815]: I0225 14:06:00.165062 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533806-l2585" Feb 25 14:06:00 crc kubenswrapper[4815]: I0225 14:06:00.168422 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 14:06:00 crc kubenswrapper[4815]: I0225 14:06:00.168797 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 14:06:00 crc kubenswrapper[4815]: I0225 14:06:00.169553 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 14:06:00 crc kubenswrapper[4815]: I0225 14:06:00.183496 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533806-l2585"] Feb 25 14:06:00 crc kubenswrapper[4815]: I0225 14:06:00.305813 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsvkd\" (UniqueName: \"kubernetes.io/projected/6269a886-839b-41f6-9c68-0759235f4ffc-kube-api-access-qsvkd\") pod \"auto-csr-approver-29533806-l2585\" (UID: \"6269a886-839b-41f6-9c68-0759235f4ffc\") " pod="openshift-infra/auto-csr-approver-29533806-l2585" Feb 25 14:06:00 crc kubenswrapper[4815]: I0225 14:06:00.407405 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsvkd\" (UniqueName: \"kubernetes.io/projected/6269a886-839b-41f6-9c68-0759235f4ffc-kube-api-access-qsvkd\") pod \"auto-csr-approver-29533806-l2585\" (UID: \"6269a886-839b-41f6-9c68-0759235f4ffc\") " pod="openshift-infra/auto-csr-approver-29533806-l2585" Feb 25 14:06:00 crc kubenswrapper[4815]: I0225 14:06:00.429554 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsvkd\" (UniqueName: \"kubernetes.io/projected/6269a886-839b-41f6-9c68-0759235f4ffc-kube-api-access-qsvkd\") pod \"auto-csr-approver-29533806-l2585\" (UID: \"6269a886-839b-41f6-9c68-0759235f4ffc\") " pod="openshift-infra/auto-csr-approver-29533806-l2585" Feb 25 14:06:00 crc kubenswrapper[4815]: I0225 14:06:00.490462 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533806-l2585" Feb 25 14:06:00 crc kubenswrapper[4815]: I0225 14:06:00.946634 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533806-l2585"] Feb 25 14:06:00 crc kubenswrapper[4815]: I0225 14:06:00.949364 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 14:06:01 crc kubenswrapper[4815]: I0225 14:06:01.622978 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533806-l2585" event={"ID":"6269a886-839b-41f6-9c68-0759235f4ffc","Type":"ContainerStarted","Data":"dd4cd88572a9bf81d7a820ef08214cc25b206bbd3aa4a726847b6709e6aa428b"} Feb 25 14:06:02 crc kubenswrapper[4815]: I0225 14:06:02.632897 4815 generic.go:334] "Generic (PLEG): container finished" podID="6269a886-839b-41f6-9c68-0759235f4ffc" containerID="a34586a0d45fed6749a23f18bd02304a1e6eedb25b5892dfa0f8043ba4597263" exitCode=0 Feb 25 14:06:02 crc kubenswrapper[4815]: I0225 14:06:02.632960 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533806-l2585" event={"ID":"6269a886-839b-41f6-9c68-0759235f4ffc","Type":"ContainerDied","Data":"a34586a0d45fed6749a23f18bd02304a1e6eedb25b5892dfa0f8043ba4597263"} Feb 25 14:06:03 crc kubenswrapper[4815]: I0225 14:06:03.998695 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533806-l2585" Feb 25 14:06:04 crc kubenswrapper[4815]: I0225 14:06:04.180555 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsvkd\" (UniqueName: \"kubernetes.io/projected/6269a886-839b-41f6-9c68-0759235f4ffc-kube-api-access-qsvkd\") pod \"6269a886-839b-41f6-9c68-0759235f4ffc\" (UID: \"6269a886-839b-41f6-9c68-0759235f4ffc\") " Feb 25 14:06:04 crc kubenswrapper[4815]: I0225 14:06:04.202925 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6269a886-839b-41f6-9c68-0759235f4ffc-kube-api-access-qsvkd" (OuterVolumeSpecName: "kube-api-access-qsvkd") pod "6269a886-839b-41f6-9c68-0759235f4ffc" (UID: "6269a886-839b-41f6-9c68-0759235f4ffc"). InnerVolumeSpecName "kube-api-access-qsvkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:06:04 crc kubenswrapper[4815]: I0225 14:06:04.282945 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsvkd\" (UniqueName: \"kubernetes.io/projected/6269a886-839b-41f6-9c68-0759235f4ffc-kube-api-access-qsvkd\") on node \"crc\" DevicePath \"\"" Feb 25 14:06:04 crc kubenswrapper[4815]: I0225 14:06:04.649786 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533806-l2585" event={"ID":"6269a886-839b-41f6-9c68-0759235f4ffc","Type":"ContainerDied","Data":"dd4cd88572a9bf81d7a820ef08214cc25b206bbd3aa4a726847b6709e6aa428b"} Feb 25 14:06:04 crc kubenswrapper[4815]: I0225 14:06:04.649846 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd4cd88572a9bf81d7a820ef08214cc25b206bbd3aa4a726847b6709e6aa428b" Feb 25 14:06:04 crc kubenswrapper[4815]: I0225 14:06:04.649894 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533806-l2585" Feb 25 14:06:05 crc kubenswrapper[4815]: I0225 14:06:05.082829 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533800-6drfm"] Feb 25 14:06:05 crc kubenswrapper[4815]: I0225 14:06:05.090268 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533800-6drfm"] Feb 25 14:06:06 crc kubenswrapper[4815]: I0225 14:06:06.946077 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46c73af1-a231-49fc-aa2e-0278b7490eb1" path="/var/lib/kubelet/pods/46c73af1-a231-49fc-aa2e-0278b7490eb1/volumes" Feb 25 14:06:21 crc kubenswrapper[4815]: I0225 14:06:21.341798 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 14:06:21 crc kubenswrapper[4815]: I0225 14:06:21.342567 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 14:06:21 crc kubenswrapper[4815]: I0225 14:06:21.342613 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 14:06:21 crc kubenswrapper[4815]: I0225 14:06:21.343401 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aa536f69b8a87db0cb31810395d99769259dc48b08c1c927d09e2d481a2a47df"} pod="openshift-machine-config-operator/machine-config-daemon-frc75" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 14:06:21 crc kubenswrapper[4815]: I0225 14:06:21.343456 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" containerID="cri-o://aa536f69b8a87db0cb31810395d99769259dc48b08c1c927d09e2d481a2a47df" gracePeriod=600 Feb 25 14:06:21 crc kubenswrapper[4815]: I0225 14:06:21.841787 4815 generic.go:334] "Generic (PLEG): container finished" podID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerID="aa536f69b8a87db0cb31810395d99769259dc48b08c1c927d09e2d481a2a47df" exitCode=0 Feb 25 14:06:21 crc kubenswrapper[4815]: I0225 14:06:21.841883 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerDied","Data":"aa536f69b8a87db0cb31810395d99769259dc48b08c1c927d09e2d481a2a47df"} Feb 25 14:06:21 crc kubenswrapper[4815]: I0225 14:06:21.842228 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerStarted","Data":"68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362"} Feb 25 14:06:21 crc kubenswrapper[4815]: I0225 14:06:21.842256 4815 scope.go:117] "RemoveContainer" containerID="9ab49ae93c9060ccc27a8e78d21e1d4989b85db07a7eddd6be30ca2d7aea5c05" Feb 25 14:06:32 crc kubenswrapper[4815]: I0225 14:06:32.824907 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sckpv"] Feb 25 14:06:32 crc kubenswrapper[4815]: E0225 14:06:32.825999 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6269a886-839b-41f6-9c68-0759235f4ffc" containerName="oc" Feb 25 14:06:32 crc kubenswrapper[4815]: I0225 14:06:32.826019 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6269a886-839b-41f6-9c68-0759235f4ffc" containerName="oc" Feb 25 14:06:32 crc kubenswrapper[4815]: I0225 14:06:32.826273 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="6269a886-839b-41f6-9c68-0759235f4ffc" containerName="oc" Feb 25 14:06:32 crc kubenswrapper[4815]: I0225 14:06:32.827975 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sckpv" Feb 25 14:06:32 crc kubenswrapper[4815]: I0225 14:06:32.837654 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sckpv"] Feb 25 14:06:32 crc kubenswrapper[4815]: I0225 14:06:32.973352 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/948321c9-1249-473a-a3bd-83819bc77682-catalog-content\") pod \"redhat-operators-sckpv\" (UID: \"948321c9-1249-473a-a3bd-83819bc77682\") " pod="openshift-marketplace/redhat-operators-sckpv" Feb 25 14:06:32 crc kubenswrapper[4815]: I0225 14:06:32.974021 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9dvz\" (UniqueName: \"kubernetes.io/projected/948321c9-1249-473a-a3bd-83819bc77682-kube-api-access-h9dvz\") pod \"redhat-operators-sckpv\" (UID: \"948321c9-1249-473a-a3bd-83819bc77682\") " pod="openshift-marketplace/redhat-operators-sckpv" Feb 25 14:06:32 crc kubenswrapper[4815]: I0225 14:06:32.974245 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/948321c9-1249-473a-a3bd-83819bc77682-utilities\") pod \"redhat-operators-sckpv\" (UID: \"948321c9-1249-473a-a3bd-83819bc77682\") " pod="openshift-marketplace/redhat-operators-sckpv" Feb 25 14:06:33 crc kubenswrapper[4815]: I0225 14:06:33.075837 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/948321c9-1249-473a-a3bd-83819bc77682-catalog-content\") pod \"redhat-operators-sckpv\" (UID: \"948321c9-1249-473a-a3bd-83819bc77682\") " pod="openshift-marketplace/redhat-operators-sckpv" Feb 25 14:06:33 crc kubenswrapper[4815]: I0225 14:06:33.075986 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9dvz\" (UniqueName: \"kubernetes.io/projected/948321c9-1249-473a-a3bd-83819bc77682-kube-api-access-h9dvz\") pod \"redhat-operators-sckpv\" (UID: \"948321c9-1249-473a-a3bd-83819bc77682\") " pod="openshift-marketplace/redhat-operators-sckpv" Feb 25 14:06:33 crc kubenswrapper[4815]: I0225 14:06:33.076039 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/948321c9-1249-473a-a3bd-83819bc77682-utilities\") pod \"redhat-operators-sckpv\" (UID: \"948321c9-1249-473a-a3bd-83819bc77682\") " pod="openshift-marketplace/redhat-operators-sckpv" Feb 25 14:06:33 crc kubenswrapper[4815]: I0225 14:06:33.076541 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/948321c9-1249-473a-a3bd-83819bc77682-utilities\") pod \"redhat-operators-sckpv\" (UID: \"948321c9-1249-473a-a3bd-83819bc77682\") " pod="openshift-marketplace/redhat-operators-sckpv" Feb 25 14:06:33 crc kubenswrapper[4815]: I0225 14:06:33.076539 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/948321c9-1249-473a-a3bd-83819bc77682-catalog-content\") pod \"redhat-operators-sckpv\" (UID: \"948321c9-1249-473a-a3bd-83819bc77682\") " pod="openshift-marketplace/redhat-operators-sckpv" Feb 25 14:06:33 crc kubenswrapper[4815]: I0225 14:06:33.107324 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9dvz\" (UniqueName: \"kubernetes.io/projected/948321c9-1249-473a-a3bd-83819bc77682-kube-api-access-h9dvz\") pod \"redhat-operators-sckpv\" (UID: \"948321c9-1249-473a-a3bd-83819bc77682\") " pod="openshift-marketplace/redhat-operators-sckpv" Feb 25 14:06:33 crc kubenswrapper[4815]: I0225 14:06:33.151068 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sckpv" Feb 25 14:06:33 crc kubenswrapper[4815]: I0225 14:06:33.626691 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sckpv"] Feb 25 14:06:33 crc kubenswrapper[4815]: I0225 14:06:33.957005 4815 generic.go:334] "Generic (PLEG): container finished" podID="948321c9-1249-473a-a3bd-83819bc77682" containerID="1d05d3dbfc94aac27f8e7a0838e0c3da025712ff6e2de405d316d688cc4303b0" exitCode=0 Feb 25 14:06:33 crc kubenswrapper[4815]: I0225 14:06:33.957046 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sckpv" event={"ID":"948321c9-1249-473a-a3bd-83819bc77682","Type":"ContainerDied","Data":"1d05d3dbfc94aac27f8e7a0838e0c3da025712ff6e2de405d316d688cc4303b0"} Feb 25 14:06:33 crc kubenswrapper[4815]: I0225 14:06:33.957072 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sckpv" event={"ID":"948321c9-1249-473a-a3bd-83819bc77682","Type":"ContainerStarted","Data":"825f8409815bf897b3af1db4b642c9a33f55e33e2f5c12ab6fe6d56fad9e092b"} Feb 25 14:06:34 crc kubenswrapper[4815]: I0225 14:06:34.968784 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sckpv" event={"ID":"948321c9-1249-473a-a3bd-83819bc77682","Type":"ContainerStarted","Data":"6b628c85afb2c8697114cbda4be4fa866c126f55e722bbc31616ffc30f784162"} Feb 25 14:06:36 crc kubenswrapper[4815]: I0225 14:06:36.988392 4815 generic.go:334] "Generic (PLEG): container finished" podID="948321c9-1249-473a-a3bd-83819bc77682" containerID="6b628c85afb2c8697114cbda4be4fa866c126f55e722bbc31616ffc30f784162" exitCode=0 Feb 25 14:06:36 crc kubenswrapper[4815]: I0225 14:06:36.988465 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sckpv" event={"ID":"948321c9-1249-473a-a3bd-83819bc77682","Type":"ContainerDied","Data":"6b628c85afb2c8697114cbda4be4fa866c126f55e722bbc31616ffc30f784162"} Feb 25 14:06:38 crc kubenswrapper[4815]: I0225 14:06:38.001720 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sckpv" event={"ID":"948321c9-1249-473a-a3bd-83819bc77682","Type":"ContainerStarted","Data":"2c7e5a264d796840c0f423ecf3e71ed722de606655efc8e4d1012d2a21bb365c"} Feb 25 14:06:38 crc kubenswrapper[4815]: I0225 14:06:38.024221 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sckpv" podStartSLOduration=2.584247751 podStartE2EDuration="6.024203444s" podCreationTimestamp="2026-02-25 14:06:32 +0000 UTC" firstStartedPulling="2026-02-25 14:06:33.958915879 +0000 UTC m=+2751.760013933" lastFinishedPulling="2026-02-25 14:06:37.398871572 +0000 UTC m=+2755.199969626" observedRunningTime="2026-02-25 14:06:38.020590971 +0000 UTC m=+2755.821689025" watchObservedRunningTime="2026-02-25 14:06:38.024203444 +0000 UTC m=+2755.825301498" Feb 25 14:06:43 crc kubenswrapper[4815]: I0225 14:06:43.151763 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sckpv" Feb 25 14:06:43 crc kubenswrapper[4815]: I0225 14:06:43.152406 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sckpv" Feb 25 14:06:44 crc kubenswrapper[4815]: I0225 14:06:44.196301 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sckpv" podUID="948321c9-1249-473a-a3bd-83819bc77682" containerName="registry-server" probeResult="failure" output=< Feb 25 14:06:44 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Feb 25 14:06:44 crc kubenswrapper[4815]: > Feb 25 14:06:49 crc kubenswrapper[4815]: I0225 14:06:49.096586 4815 scope.go:117] "RemoveContainer" containerID="6ca4da694e28152b3541b51daf3321e9636dddf07bee483b236ebb5c0509eadb" Feb 25 14:06:53 crc kubenswrapper[4815]: I0225 14:06:53.233641 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sckpv" Feb 25 14:06:53 crc kubenswrapper[4815]: I0225 14:06:53.294479 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sckpv" Feb 25 14:06:53 crc kubenswrapper[4815]: I0225 14:06:53.492744 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sckpv"] Feb 25 14:06:55 crc kubenswrapper[4815]: I0225 14:06:55.172017 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sckpv" podUID="948321c9-1249-473a-a3bd-83819bc77682" containerName="registry-server" containerID="cri-o://2c7e5a264d796840c0f423ecf3e71ed722de606655efc8e4d1012d2a21bb365c" gracePeriod=2 Feb 25 14:06:55 crc kubenswrapper[4815]: I0225 14:06:55.756024 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sckpv" Feb 25 14:06:55 crc kubenswrapper[4815]: I0225 14:06:55.835664 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/948321c9-1249-473a-a3bd-83819bc77682-utilities\") pod \"948321c9-1249-473a-a3bd-83819bc77682\" (UID: \"948321c9-1249-473a-a3bd-83819bc77682\") " Feb 25 14:06:55 crc kubenswrapper[4815]: I0225 14:06:55.835772 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9dvz\" (UniqueName: \"kubernetes.io/projected/948321c9-1249-473a-a3bd-83819bc77682-kube-api-access-h9dvz\") pod \"948321c9-1249-473a-a3bd-83819bc77682\" (UID: \"948321c9-1249-473a-a3bd-83819bc77682\") " Feb 25 14:06:55 crc kubenswrapper[4815]: I0225 14:06:55.835911 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/948321c9-1249-473a-a3bd-83819bc77682-catalog-content\") pod \"948321c9-1249-473a-a3bd-83819bc77682\" (UID: \"948321c9-1249-473a-a3bd-83819bc77682\") " Feb 25 14:06:55 crc kubenswrapper[4815]: I0225 14:06:55.840069 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/948321c9-1249-473a-a3bd-83819bc77682-utilities" (OuterVolumeSpecName: "utilities") pod "948321c9-1249-473a-a3bd-83819bc77682" (UID: "948321c9-1249-473a-a3bd-83819bc77682"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:06:55 crc kubenswrapper[4815]: I0225 14:06:55.846750 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/948321c9-1249-473a-a3bd-83819bc77682-kube-api-access-h9dvz" (OuterVolumeSpecName: "kube-api-access-h9dvz") pod "948321c9-1249-473a-a3bd-83819bc77682" (UID: "948321c9-1249-473a-a3bd-83819bc77682"). InnerVolumeSpecName "kube-api-access-h9dvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:06:55 crc kubenswrapper[4815]: I0225 14:06:55.938154 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/948321c9-1249-473a-a3bd-83819bc77682-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 14:06:55 crc kubenswrapper[4815]: I0225 14:06:55.938192 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9dvz\" (UniqueName: \"kubernetes.io/projected/948321c9-1249-473a-a3bd-83819bc77682-kube-api-access-h9dvz\") on node \"crc\" DevicePath \"\"" Feb 25 14:06:55 crc kubenswrapper[4815]: I0225 14:06:55.968983 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/948321c9-1249-473a-a3bd-83819bc77682-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "948321c9-1249-473a-a3bd-83819bc77682" (UID: "948321c9-1249-473a-a3bd-83819bc77682"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:06:56 crc kubenswrapper[4815]: I0225 14:06:56.039671 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/948321c9-1249-473a-a3bd-83819bc77682-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 14:06:56 crc kubenswrapper[4815]: I0225 14:06:56.183148 4815 generic.go:334] "Generic (PLEG): container finished" podID="948321c9-1249-473a-a3bd-83819bc77682" containerID="2c7e5a264d796840c0f423ecf3e71ed722de606655efc8e4d1012d2a21bb365c" exitCode=0 Feb 25 14:06:56 crc kubenswrapper[4815]: I0225 14:06:56.183196 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sckpv" event={"ID":"948321c9-1249-473a-a3bd-83819bc77682","Type":"ContainerDied","Data":"2c7e5a264d796840c0f423ecf3e71ed722de606655efc8e4d1012d2a21bb365c"} Feb 25 14:06:56 crc kubenswrapper[4815]: I0225 14:06:56.183227 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sckpv" event={"ID":"948321c9-1249-473a-a3bd-83819bc77682","Type":"ContainerDied","Data":"825f8409815bf897b3af1db4b642c9a33f55e33e2f5c12ab6fe6d56fad9e092b"} Feb 25 14:06:56 crc kubenswrapper[4815]: I0225 14:06:56.183246 4815 scope.go:117] "RemoveContainer" containerID="2c7e5a264d796840c0f423ecf3e71ed722de606655efc8e4d1012d2a21bb365c" Feb 25 14:06:56 crc kubenswrapper[4815]: I0225 14:06:56.183372 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sckpv" Feb 25 14:06:56 crc kubenswrapper[4815]: I0225 14:06:56.221544 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sckpv"] Feb 25 14:06:56 crc kubenswrapper[4815]: I0225 14:06:56.222939 4815 scope.go:117] "RemoveContainer" containerID="6b628c85afb2c8697114cbda4be4fa866c126f55e722bbc31616ffc30f784162" Feb 25 14:06:56 crc kubenswrapper[4815]: I0225 14:06:56.230878 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sckpv"] Feb 25 14:06:56 crc kubenswrapper[4815]: I0225 14:06:56.252419 4815 scope.go:117] "RemoveContainer" containerID="1d05d3dbfc94aac27f8e7a0838e0c3da025712ff6e2de405d316d688cc4303b0" Feb 25 14:06:56 crc kubenswrapper[4815]: I0225 14:06:56.293160 4815 scope.go:117] "RemoveContainer" containerID="2c7e5a264d796840c0f423ecf3e71ed722de606655efc8e4d1012d2a21bb365c" Feb 25 14:06:56 crc kubenswrapper[4815]: E0225 14:06:56.293546 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c7e5a264d796840c0f423ecf3e71ed722de606655efc8e4d1012d2a21bb365c\": container with ID starting with 2c7e5a264d796840c0f423ecf3e71ed722de606655efc8e4d1012d2a21bb365c not found: ID does not exist" containerID="2c7e5a264d796840c0f423ecf3e71ed722de606655efc8e4d1012d2a21bb365c" Feb 25 14:06:56 crc kubenswrapper[4815]: I0225 14:06:56.293582 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c7e5a264d796840c0f423ecf3e71ed722de606655efc8e4d1012d2a21bb365c"} err="failed to get container status \"2c7e5a264d796840c0f423ecf3e71ed722de606655efc8e4d1012d2a21bb365c\": rpc error: code = NotFound desc = could not find container \"2c7e5a264d796840c0f423ecf3e71ed722de606655efc8e4d1012d2a21bb365c\": container with ID starting with 2c7e5a264d796840c0f423ecf3e71ed722de606655efc8e4d1012d2a21bb365c not found: ID does not exist" Feb 25 14:06:56 crc kubenswrapper[4815]: I0225 14:06:56.293603 4815 scope.go:117] "RemoveContainer" containerID="6b628c85afb2c8697114cbda4be4fa866c126f55e722bbc31616ffc30f784162" Feb 25 14:06:56 crc kubenswrapper[4815]: E0225 14:06:56.293864 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b628c85afb2c8697114cbda4be4fa866c126f55e722bbc31616ffc30f784162\": container with ID starting with 6b628c85afb2c8697114cbda4be4fa866c126f55e722bbc31616ffc30f784162 not found: ID does not exist" containerID="6b628c85afb2c8697114cbda4be4fa866c126f55e722bbc31616ffc30f784162" Feb 25 14:06:56 crc kubenswrapper[4815]: I0225 14:06:56.293889 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b628c85afb2c8697114cbda4be4fa866c126f55e722bbc31616ffc30f784162"} err="failed to get container status \"6b628c85afb2c8697114cbda4be4fa866c126f55e722bbc31616ffc30f784162\": rpc error: code = NotFound desc = could not find container \"6b628c85afb2c8697114cbda4be4fa866c126f55e722bbc31616ffc30f784162\": container with ID starting with 6b628c85afb2c8697114cbda4be4fa866c126f55e722bbc31616ffc30f784162 not found: ID does not exist" Feb 25 14:06:56 crc kubenswrapper[4815]: I0225 14:06:56.293908 4815 scope.go:117] "RemoveContainer" containerID="1d05d3dbfc94aac27f8e7a0838e0c3da025712ff6e2de405d316d688cc4303b0" Feb 25 14:06:56 crc kubenswrapper[4815]: E0225 14:06:56.294502 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d05d3dbfc94aac27f8e7a0838e0c3da025712ff6e2de405d316d688cc4303b0\": container with ID starting with 1d05d3dbfc94aac27f8e7a0838e0c3da025712ff6e2de405d316d688cc4303b0 not found: ID does not exist" containerID="1d05d3dbfc94aac27f8e7a0838e0c3da025712ff6e2de405d316d688cc4303b0" Feb 25 14:06:56 crc kubenswrapper[4815]: I0225 14:06:56.294587 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d05d3dbfc94aac27f8e7a0838e0c3da025712ff6e2de405d316d688cc4303b0"} err="failed to get container status \"1d05d3dbfc94aac27f8e7a0838e0c3da025712ff6e2de405d316d688cc4303b0\": rpc error: code = NotFound desc = could not find container \"1d05d3dbfc94aac27f8e7a0838e0c3da025712ff6e2de405d316d688cc4303b0\": container with ID starting with 1d05d3dbfc94aac27f8e7a0838e0c3da025712ff6e2de405d316d688cc4303b0 not found: ID does not exist" Feb 25 14:06:56 crc kubenswrapper[4815]: I0225 14:06:56.955845 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="948321c9-1249-473a-a3bd-83819bc77682" path="/var/lib/kubelet/pods/948321c9-1249-473a-a3bd-83819bc77682/volumes" Feb 25 14:08:00 crc kubenswrapper[4815]: I0225 14:08:00.144242 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533808-nlxhc"] Feb 25 14:08:00 crc kubenswrapper[4815]: E0225 14:08:00.145162 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="948321c9-1249-473a-a3bd-83819bc77682" containerName="extract-utilities" Feb 25 14:08:00 crc kubenswrapper[4815]: I0225 14:08:00.145174 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="948321c9-1249-473a-a3bd-83819bc77682" containerName="extract-utilities" Feb 25 14:08:00 crc kubenswrapper[4815]: E0225 14:08:00.145205 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="948321c9-1249-473a-a3bd-83819bc77682" containerName="extract-content" Feb 25 14:08:00 crc kubenswrapper[4815]: I0225 14:08:00.145213 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="948321c9-1249-473a-a3bd-83819bc77682" containerName="extract-content" Feb 25 14:08:00 crc kubenswrapper[4815]: E0225 14:08:00.145259 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="948321c9-1249-473a-a3bd-83819bc77682" containerName="registry-server" Feb 25 14:08:00 crc kubenswrapper[4815]: I0225 14:08:00.145284 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="948321c9-1249-473a-a3bd-83819bc77682" containerName="registry-server" Feb 25 14:08:00 crc kubenswrapper[4815]: I0225 14:08:00.145460 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="948321c9-1249-473a-a3bd-83819bc77682" containerName="registry-server" Feb 25 14:08:00 crc kubenswrapper[4815]: I0225 14:08:00.146077 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533808-nlxhc" Feb 25 14:08:00 crc kubenswrapper[4815]: I0225 14:08:00.148195 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 14:08:00 crc kubenswrapper[4815]: I0225 14:08:00.149857 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 14:08:00 crc kubenswrapper[4815]: I0225 14:08:00.150261 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 14:08:00 crc kubenswrapper[4815]: I0225 14:08:00.176639 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533808-nlxhc"] Feb 25 14:08:00 crc kubenswrapper[4815]: I0225 14:08:00.324597 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsbhz\" (UniqueName: \"kubernetes.io/projected/90489231-445c-4533-833e-3c0116ff824e-kube-api-access-jsbhz\") pod \"auto-csr-approver-29533808-nlxhc\" (UID: \"90489231-445c-4533-833e-3c0116ff824e\") " pod="openshift-infra/auto-csr-approver-29533808-nlxhc" Feb 25 14:08:00 crc kubenswrapper[4815]: I0225 14:08:00.426761 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsbhz\" (UniqueName: \"kubernetes.io/projected/90489231-445c-4533-833e-3c0116ff824e-kube-api-access-jsbhz\") pod \"auto-csr-approver-29533808-nlxhc\" (UID: \"90489231-445c-4533-833e-3c0116ff824e\") " pod="openshift-infra/auto-csr-approver-29533808-nlxhc" Feb 25 14:08:00 crc kubenswrapper[4815]: I0225 14:08:00.455863 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsbhz\" (UniqueName: \"kubernetes.io/projected/90489231-445c-4533-833e-3c0116ff824e-kube-api-access-jsbhz\") pod \"auto-csr-approver-29533808-nlxhc\" (UID: \"90489231-445c-4533-833e-3c0116ff824e\") " pod="openshift-infra/auto-csr-approver-29533808-nlxhc" Feb 25 14:08:00 crc kubenswrapper[4815]: I0225 14:08:00.465707 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533808-nlxhc" Feb 25 14:08:00 crc kubenswrapper[4815]: I0225 14:08:00.946672 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533808-nlxhc"] Feb 25 14:08:01 crc kubenswrapper[4815]: I0225 14:08:01.835430 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533808-nlxhc" event={"ID":"90489231-445c-4533-833e-3c0116ff824e","Type":"ContainerStarted","Data":"1d9638e7bee6d8329d0fece7153a6fee494ec45057ad102b963cf15d3b018a9b"} Feb 25 14:08:02 crc kubenswrapper[4815]: I0225 14:08:02.847723 4815 generic.go:334] "Generic (PLEG): container finished" podID="90489231-445c-4533-833e-3c0116ff824e" containerID="3bc65c3a29af0bc97a6740353fa892252c3148b0388e77701f0b4bb5beca3925" exitCode=0 Feb 25 14:08:02 crc kubenswrapper[4815]: I0225 14:08:02.848170 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533808-nlxhc" event={"ID":"90489231-445c-4533-833e-3c0116ff824e","Type":"ContainerDied","Data":"3bc65c3a29af0bc97a6740353fa892252c3148b0388e77701f0b4bb5beca3925"} Feb 25 14:08:04 crc kubenswrapper[4815]: I0225 14:08:04.237437 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533808-nlxhc" Feb 25 14:08:04 crc kubenswrapper[4815]: I0225 14:08:04.298102 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsbhz\" (UniqueName: \"kubernetes.io/projected/90489231-445c-4533-833e-3c0116ff824e-kube-api-access-jsbhz\") pod \"90489231-445c-4533-833e-3c0116ff824e\" (UID: \"90489231-445c-4533-833e-3c0116ff824e\") " Feb 25 14:08:04 crc kubenswrapper[4815]: I0225 14:08:04.304839 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90489231-445c-4533-833e-3c0116ff824e-kube-api-access-jsbhz" (OuterVolumeSpecName: "kube-api-access-jsbhz") pod "90489231-445c-4533-833e-3c0116ff824e" (UID: "90489231-445c-4533-833e-3c0116ff824e"). InnerVolumeSpecName "kube-api-access-jsbhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:08:04 crc kubenswrapper[4815]: I0225 14:08:04.399787 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsbhz\" (UniqueName: \"kubernetes.io/projected/90489231-445c-4533-833e-3c0116ff824e-kube-api-access-jsbhz\") on node \"crc\" DevicePath \"\"" Feb 25 14:08:04 crc kubenswrapper[4815]: I0225 14:08:04.872165 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533808-nlxhc" event={"ID":"90489231-445c-4533-833e-3c0116ff824e","Type":"ContainerDied","Data":"1d9638e7bee6d8329d0fece7153a6fee494ec45057ad102b963cf15d3b018a9b"} Feb 25 14:08:04 crc kubenswrapper[4815]: I0225 14:08:04.872222 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d9638e7bee6d8329d0fece7153a6fee494ec45057ad102b963cf15d3b018a9b" Feb 25 14:08:04 crc kubenswrapper[4815]: I0225 14:08:04.872229 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533808-nlxhc" Feb 25 14:08:05 crc kubenswrapper[4815]: I0225 14:08:05.319668 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533802-mkd6f"] Feb 25 14:08:05 crc kubenswrapper[4815]: I0225 14:08:05.333898 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533802-mkd6f"] Feb 25 14:08:06 crc kubenswrapper[4815]: I0225 14:08:06.960807 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e236e592-fc11-48e8-a1d4-65178884f48a" path="/var/lib/kubelet/pods/e236e592-fc11-48e8-a1d4-65178884f48a/volumes" Feb 25 14:08:21 crc kubenswrapper[4815]: I0225 14:08:21.342797 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 14:08:21 crc kubenswrapper[4815]: I0225 14:08:21.343669 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 14:08:49 crc kubenswrapper[4815]: I0225 14:08:49.224438 4815 scope.go:117] "RemoveContainer" containerID="111efb696e991623cb96fb6d345e3d2aeb0175e78ffe239cfddec2911985a3d2" Feb 25 14:08:51 crc kubenswrapper[4815]: I0225 14:08:51.342024 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 14:08:51 crc kubenswrapper[4815]: I0225 14:08:51.342438 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 14:09:21 crc kubenswrapper[4815]: I0225 14:09:21.342336 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 14:09:21 crc kubenswrapper[4815]: I0225 14:09:21.343781 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 14:09:21 crc kubenswrapper[4815]: I0225 14:09:21.343884 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 14:09:21 crc kubenswrapper[4815]: I0225 14:09:21.345172 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362"} pod="openshift-machine-config-operator/machine-config-daemon-frc75" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 14:09:21 crc kubenswrapper[4815]: I0225 14:09:21.345255 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" containerID="cri-o://68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" gracePeriod=600 Feb 25 14:09:21 crc kubenswrapper[4815]: E0225 14:09:21.473360 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:09:21 crc kubenswrapper[4815]: I0225 14:09:21.725109 4815 generic.go:334] "Generic (PLEG): container finished" podID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" exitCode=0 Feb 25 14:09:21 crc kubenswrapper[4815]: I0225 14:09:21.725163 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerDied","Data":"68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362"} Feb 25 14:09:21 crc kubenswrapper[4815]: I0225 14:09:21.725196 4815 scope.go:117] "RemoveContainer" containerID="aa536f69b8a87db0cb31810395d99769259dc48b08c1c927d09e2d481a2a47df" Feb 25 14:09:21 crc kubenswrapper[4815]: I0225 14:09:21.725848 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:09:21 crc kubenswrapper[4815]: E0225 14:09:21.726222 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:09:32 crc kubenswrapper[4815]: I0225 14:09:32.944363 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:09:32 crc kubenswrapper[4815]: E0225 14:09:32.945245 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:09:46 crc kubenswrapper[4815]: I0225 14:09:46.935603 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:09:46 crc kubenswrapper[4815]: E0225 14:09:46.936399 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:10:00 crc kubenswrapper[4815]: I0225 14:10:00.157437 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533810-j9ssf"] Feb 25 14:10:00 crc kubenswrapper[4815]: E0225 14:10:00.159162 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90489231-445c-4533-833e-3c0116ff824e" containerName="oc" Feb 25 14:10:00 crc kubenswrapper[4815]: I0225 14:10:00.159190 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="90489231-445c-4533-833e-3c0116ff824e" containerName="oc" Feb 25 14:10:00 crc kubenswrapper[4815]: I0225 14:10:00.159618 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="90489231-445c-4533-833e-3c0116ff824e" containerName="oc" Feb 25 14:10:00 crc kubenswrapper[4815]: I0225 14:10:00.160880 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533810-j9ssf" Feb 25 14:10:00 crc kubenswrapper[4815]: I0225 14:10:00.166259 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 14:10:00 crc kubenswrapper[4815]: I0225 14:10:00.168267 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 14:10:00 crc kubenswrapper[4815]: I0225 14:10:00.168405 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 14:10:00 crc kubenswrapper[4815]: I0225 14:10:00.175236 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533810-j9ssf"] Feb 25 14:10:00 crc kubenswrapper[4815]: I0225 14:10:00.319934 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxsfk\" (UniqueName: \"kubernetes.io/projected/4f5569c7-1d59-4aec-9f95-a27707e08b52-kube-api-access-fxsfk\") pod \"auto-csr-approver-29533810-j9ssf\" (UID: \"4f5569c7-1d59-4aec-9f95-a27707e08b52\") " pod="openshift-infra/auto-csr-approver-29533810-j9ssf" Feb 25 14:10:00 crc kubenswrapper[4815]: I0225 14:10:00.422679 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxsfk\" (UniqueName: \"kubernetes.io/projected/4f5569c7-1d59-4aec-9f95-a27707e08b52-kube-api-access-fxsfk\") pod \"auto-csr-approver-29533810-j9ssf\" (UID: \"4f5569c7-1d59-4aec-9f95-a27707e08b52\") " pod="openshift-infra/auto-csr-approver-29533810-j9ssf" Feb 25 14:10:00 crc kubenswrapper[4815]: I0225 14:10:00.451281 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxsfk\" (UniqueName: \"kubernetes.io/projected/4f5569c7-1d59-4aec-9f95-a27707e08b52-kube-api-access-fxsfk\") pod \"auto-csr-approver-29533810-j9ssf\" (UID: \"4f5569c7-1d59-4aec-9f95-a27707e08b52\") " pod="openshift-infra/auto-csr-approver-29533810-j9ssf" Feb 25 14:10:00 crc kubenswrapper[4815]: I0225 14:10:00.480935 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533810-j9ssf" Feb 25 14:10:00 crc kubenswrapper[4815]: I0225 14:10:00.935790 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:10:00 crc kubenswrapper[4815]: E0225 14:10:00.936292 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:10:01 crc kubenswrapper[4815]: I0225 14:10:01.002283 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533810-j9ssf"] Feb 25 14:10:01 crc kubenswrapper[4815]: I0225 14:10:01.131905 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533810-j9ssf" event={"ID":"4f5569c7-1d59-4aec-9f95-a27707e08b52","Type":"ContainerStarted","Data":"13c87cc79c31eadc91673f31ca76fcf507dde910988e2967f9371eed80653813"} Feb 25 14:10:03 crc kubenswrapper[4815]: I0225 14:10:03.161614 4815 generic.go:334] "Generic (PLEG): container finished" podID="4f5569c7-1d59-4aec-9f95-a27707e08b52" containerID="ba245b6ab9f0ba822c6ae549e58c31c8928e482602ebdbb2db4363f34083044a" exitCode=0 Feb 25 14:10:03 crc kubenswrapper[4815]: I0225 14:10:03.161673 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533810-j9ssf" event={"ID":"4f5569c7-1d59-4aec-9f95-a27707e08b52","Type":"ContainerDied","Data":"ba245b6ab9f0ba822c6ae549e58c31c8928e482602ebdbb2db4363f34083044a"} Feb 25 14:10:04 crc kubenswrapper[4815]: I0225 14:10:04.565205 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533810-j9ssf" Feb 25 14:10:04 crc kubenswrapper[4815]: I0225 14:10:04.712391 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxsfk\" (UniqueName: \"kubernetes.io/projected/4f5569c7-1d59-4aec-9f95-a27707e08b52-kube-api-access-fxsfk\") pod \"4f5569c7-1d59-4aec-9f95-a27707e08b52\" (UID: \"4f5569c7-1d59-4aec-9f95-a27707e08b52\") " Feb 25 14:10:04 crc kubenswrapper[4815]: I0225 14:10:04.722636 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f5569c7-1d59-4aec-9f95-a27707e08b52-kube-api-access-fxsfk" (OuterVolumeSpecName: "kube-api-access-fxsfk") pod "4f5569c7-1d59-4aec-9f95-a27707e08b52" (UID: "4f5569c7-1d59-4aec-9f95-a27707e08b52"). InnerVolumeSpecName "kube-api-access-fxsfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:10:04 crc kubenswrapper[4815]: I0225 14:10:04.814541 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxsfk\" (UniqueName: \"kubernetes.io/projected/4f5569c7-1d59-4aec-9f95-a27707e08b52-kube-api-access-fxsfk\") on node \"crc\" DevicePath \"\"" Feb 25 14:10:05 crc kubenswrapper[4815]: I0225 14:10:05.182358 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533810-j9ssf" event={"ID":"4f5569c7-1d59-4aec-9f95-a27707e08b52","Type":"ContainerDied","Data":"13c87cc79c31eadc91673f31ca76fcf507dde910988e2967f9371eed80653813"} Feb 25 14:10:05 crc kubenswrapper[4815]: I0225 14:10:05.182683 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13c87cc79c31eadc91673f31ca76fcf507dde910988e2967f9371eed80653813" Feb 25 14:10:05 crc kubenswrapper[4815]: I0225 14:10:05.182689 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533810-j9ssf" Feb 25 14:10:05 crc kubenswrapper[4815]: I0225 14:10:05.638678 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533804-rpvkq"] Feb 25 14:10:05 crc kubenswrapper[4815]: I0225 14:10:05.645444 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533804-rpvkq"] Feb 25 14:10:06 crc kubenswrapper[4815]: I0225 14:10:06.948730 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d" path="/var/lib/kubelet/pods/8fa0c6ba-5bb6-4fcd-88d0-3c572a4c761d/volumes" Feb 25 14:10:13 crc kubenswrapper[4815]: I0225 14:10:13.935071 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:10:13 crc kubenswrapper[4815]: E0225 14:10:13.935868 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:10:26 crc kubenswrapper[4815]: I0225 14:10:26.936035 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:10:26 crc kubenswrapper[4815]: E0225 14:10:26.936602 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:10:41 crc kubenswrapper[4815]: I0225 14:10:41.936659 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:10:41 crc kubenswrapper[4815]: E0225 14:10:41.937669 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:10:49 crc kubenswrapper[4815]: I0225 14:10:49.342891 4815 scope.go:117] "RemoveContainer" containerID="f64b6650bc4aaa76893d99c1eefaa798e69267d5d795d21dbfed923c703e4c9a" Feb 25 14:10:52 crc kubenswrapper[4815]: I0225 14:10:52.941869 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:10:52 crc kubenswrapper[4815]: E0225 14:10:52.942635 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:11:03 crc kubenswrapper[4815]: I0225 14:11:03.936001 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:11:03 crc kubenswrapper[4815]: E0225 14:11:03.936667 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:11:15 crc kubenswrapper[4815]: I0225 14:11:15.935812 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:11:15 crc kubenswrapper[4815]: E0225 14:11:15.936557 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:11:16 crc kubenswrapper[4815]: I0225 14:11:16.502951 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jdmp9"] Feb 25 14:11:16 crc kubenswrapper[4815]: E0225 14:11:16.504798 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f5569c7-1d59-4aec-9f95-a27707e08b52" containerName="oc" Feb 25 14:11:16 crc kubenswrapper[4815]: I0225 14:11:16.504832 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f5569c7-1d59-4aec-9f95-a27707e08b52" containerName="oc" Feb 25 14:11:16 crc kubenswrapper[4815]: I0225 14:11:16.505604 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f5569c7-1d59-4aec-9f95-a27707e08b52" containerName="oc" Feb 25 14:11:16 crc kubenswrapper[4815]: I0225 14:11:16.509330 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jdmp9" Feb 25 14:11:16 crc kubenswrapper[4815]: I0225 14:11:16.559043 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jdmp9"] Feb 25 14:11:16 crc kubenswrapper[4815]: I0225 14:11:16.570127 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccaa1a41-0180-48a5-b835-dfcbe76061fb-catalog-content\") pod \"certified-operators-jdmp9\" (UID: \"ccaa1a41-0180-48a5-b835-dfcbe76061fb\") " pod="openshift-marketplace/certified-operators-jdmp9" Feb 25 14:11:16 crc kubenswrapper[4815]: I0225 14:11:16.570459 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccaa1a41-0180-48a5-b835-dfcbe76061fb-utilities\") pod \"certified-operators-jdmp9\" (UID: \"ccaa1a41-0180-48a5-b835-dfcbe76061fb\") " pod="openshift-marketplace/certified-operators-jdmp9" Feb 25 14:11:16 crc kubenswrapper[4815]: I0225 14:11:16.570559 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8ngh\" (UniqueName: \"kubernetes.io/projected/ccaa1a41-0180-48a5-b835-dfcbe76061fb-kube-api-access-x8ngh\") pod \"certified-operators-jdmp9\" (UID: \"ccaa1a41-0180-48a5-b835-dfcbe76061fb\") " pod="openshift-marketplace/certified-operators-jdmp9" Feb 25 14:11:16 crc kubenswrapper[4815]: I0225 14:11:16.671918 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccaa1a41-0180-48a5-b835-dfcbe76061fb-catalog-content\") pod \"certified-operators-jdmp9\" (UID: \"ccaa1a41-0180-48a5-b835-dfcbe76061fb\") " pod="openshift-marketplace/certified-operators-jdmp9" Feb 25 14:11:16 crc kubenswrapper[4815]: I0225 14:11:16.671964 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccaa1a41-0180-48a5-b835-dfcbe76061fb-utilities\") pod \"certified-operators-jdmp9\" (UID: \"ccaa1a41-0180-48a5-b835-dfcbe76061fb\") " pod="openshift-marketplace/certified-operators-jdmp9" Feb 25 14:11:16 crc kubenswrapper[4815]: I0225 14:11:16.672028 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8ngh\" (UniqueName: \"kubernetes.io/projected/ccaa1a41-0180-48a5-b835-dfcbe76061fb-kube-api-access-x8ngh\") pod \"certified-operators-jdmp9\" (UID: \"ccaa1a41-0180-48a5-b835-dfcbe76061fb\") " pod="openshift-marketplace/certified-operators-jdmp9" Feb 25 14:11:16 crc kubenswrapper[4815]: I0225 14:11:16.672444 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccaa1a41-0180-48a5-b835-dfcbe76061fb-utilities\") pod \"certified-operators-jdmp9\" (UID: \"ccaa1a41-0180-48a5-b835-dfcbe76061fb\") " pod="openshift-marketplace/certified-operators-jdmp9" Feb 25 14:11:16 crc kubenswrapper[4815]: I0225 14:11:16.672726 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccaa1a41-0180-48a5-b835-dfcbe76061fb-catalog-content\") pod \"certified-operators-jdmp9\" (UID: \"ccaa1a41-0180-48a5-b835-dfcbe76061fb\") " pod="openshift-marketplace/certified-operators-jdmp9" Feb 25 14:11:16 crc kubenswrapper[4815]: I0225 14:11:16.696899 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8ngh\" (UniqueName: \"kubernetes.io/projected/ccaa1a41-0180-48a5-b835-dfcbe76061fb-kube-api-access-x8ngh\") pod \"certified-operators-jdmp9\" (UID: \"ccaa1a41-0180-48a5-b835-dfcbe76061fb\") " pod="openshift-marketplace/certified-operators-jdmp9" Feb 25 14:11:16 crc kubenswrapper[4815]: I0225 14:11:16.830629 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jdmp9" Feb 25 14:11:17 crc kubenswrapper[4815]: I0225 14:11:17.296729 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jdmp9"] Feb 25 14:11:17 crc kubenswrapper[4815]: I0225 14:11:17.905621 4815 generic.go:334] "Generic (PLEG): container finished" podID="ccaa1a41-0180-48a5-b835-dfcbe76061fb" containerID="39ae088bc8ae38d3a886d4515eea7560ac9eadc2eb080b93ac78b7f135bcbdba" exitCode=0 Feb 25 14:11:17 crc kubenswrapper[4815]: I0225 14:11:17.905792 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jdmp9" event={"ID":"ccaa1a41-0180-48a5-b835-dfcbe76061fb","Type":"ContainerDied","Data":"39ae088bc8ae38d3a886d4515eea7560ac9eadc2eb080b93ac78b7f135bcbdba"} Feb 25 14:11:17 crc kubenswrapper[4815]: I0225 14:11:17.905948 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jdmp9" event={"ID":"ccaa1a41-0180-48a5-b835-dfcbe76061fb","Type":"ContainerStarted","Data":"3b6e70807dea6a024b550b2b5ea3679fd52ed8338d3b979892423ed887e012e0"} Feb 25 14:11:17 crc kubenswrapper[4815]: I0225 14:11:17.908755 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 14:11:18 crc kubenswrapper[4815]: I0225 14:11:18.916035 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jdmp9" event={"ID":"ccaa1a41-0180-48a5-b835-dfcbe76061fb","Type":"ContainerStarted","Data":"0a1c3c13699b75ac9a14a38ecb32e7674c4820d009a5ef747c474a80986afd8d"} Feb 25 14:11:20 crc kubenswrapper[4815]: I0225 14:11:20.942212 4815 generic.go:334] "Generic (PLEG): container finished" podID="ccaa1a41-0180-48a5-b835-dfcbe76061fb" containerID="0a1c3c13699b75ac9a14a38ecb32e7674c4820d009a5ef747c474a80986afd8d" exitCode=0 Feb 25 14:11:20 crc kubenswrapper[4815]: I0225 14:11:20.950993 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jdmp9" event={"ID":"ccaa1a41-0180-48a5-b835-dfcbe76061fb","Type":"ContainerDied","Data":"0a1c3c13699b75ac9a14a38ecb32e7674c4820d009a5ef747c474a80986afd8d"} Feb 25 14:11:21 crc kubenswrapper[4815]: I0225 14:11:21.959920 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jdmp9" event={"ID":"ccaa1a41-0180-48a5-b835-dfcbe76061fb","Type":"ContainerStarted","Data":"5d15df0c5f3622a0b870057ae6af738083bfd3c2d9f20f98143d99152a52351f"} Feb 25 14:11:21 crc kubenswrapper[4815]: I0225 14:11:21.988941 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jdmp9" podStartSLOduration=2.51815482 podStartE2EDuration="5.988913899s" podCreationTimestamp="2026-02-25 14:11:16 +0000 UTC" firstStartedPulling="2026-02-25 14:11:17.908368914 +0000 UTC m=+3035.709466988" lastFinishedPulling="2026-02-25 14:11:21.379128013 +0000 UTC m=+3039.180226067" observedRunningTime="2026-02-25 14:11:21.97905015 +0000 UTC m=+3039.780148294" watchObservedRunningTime="2026-02-25 14:11:21.988913899 +0000 UTC m=+3039.790011993" Feb 25 14:11:26 crc kubenswrapper[4815]: I0225 14:11:26.830829 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jdmp9" Feb 25 14:11:26 crc kubenswrapper[4815]: I0225 14:11:26.831550 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jdmp9" Feb 25 14:11:26 crc kubenswrapper[4815]: I0225 14:11:26.893963 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jdmp9" Feb 25 14:11:27 crc kubenswrapper[4815]: I0225 14:11:27.071189 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jdmp9" Feb 25 14:11:27 crc kubenswrapper[4815]: I0225 14:11:27.136605 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jdmp9"] Feb 25 14:11:28 crc kubenswrapper[4815]: I0225 14:11:28.938429 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:11:28 crc kubenswrapper[4815]: E0225 14:11:28.941163 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:11:29 crc kubenswrapper[4815]: I0225 14:11:29.031067 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jdmp9" podUID="ccaa1a41-0180-48a5-b835-dfcbe76061fb" containerName="registry-server" containerID="cri-o://5d15df0c5f3622a0b870057ae6af738083bfd3c2d9f20f98143d99152a52351f" gracePeriod=2 Feb 25 14:11:29 crc kubenswrapper[4815]: I0225 14:11:29.630274 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jdmp9" Feb 25 14:11:29 crc kubenswrapper[4815]: I0225 14:11:29.786671 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccaa1a41-0180-48a5-b835-dfcbe76061fb-utilities\") pod \"ccaa1a41-0180-48a5-b835-dfcbe76061fb\" (UID: \"ccaa1a41-0180-48a5-b835-dfcbe76061fb\") " Feb 25 14:11:29 crc kubenswrapper[4815]: I0225 14:11:29.787227 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccaa1a41-0180-48a5-b835-dfcbe76061fb-catalog-content\") pod \"ccaa1a41-0180-48a5-b835-dfcbe76061fb\" (UID: \"ccaa1a41-0180-48a5-b835-dfcbe76061fb\") " Feb 25 14:11:29 crc kubenswrapper[4815]: I0225 14:11:29.787357 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8ngh\" (UniqueName: \"kubernetes.io/projected/ccaa1a41-0180-48a5-b835-dfcbe76061fb-kube-api-access-x8ngh\") pod \"ccaa1a41-0180-48a5-b835-dfcbe76061fb\" (UID: \"ccaa1a41-0180-48a5-b835-dfcbe76061fb\") " Feb 25 14:11:29 crc kubenswrapper[4815]: I0225 14:11:29.787684 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccaa1a41-0180-48a5-b835-dfcbe76061fb-utilities" (OuterVolumeSpecName: "utilities") pod "ccaa1a41-0180-48a5-b835-dfcbe76061fb" (UID: "ccaa1a41-0180-48a5-b835-dfcbe76061fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:11:29 crc kubenswrapper[4815]: I0225 14:11:29.788525 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccaa1a41-0180-48a5-b835-dfcbe76061fb-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 14:11:29 crc kubenswrapper[4815]: I0225 14:11:29.794848 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccaa1a41-0180-48a5-b835-dfcbe76061fb-kube-api-access-x8ngh" (OuterVolumeSpecName: "kube-api-access-x8ngh") pod "ccaa1a41-0180-48a5-b835-dfcbe76061fb" (UID: "ccaa1a41-0180-48a5-b835-dfcbe76061fb"). InnerVolumeSpecName "kube-api-access-x8ngh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:11:29 crc kubenswrapper[4815]: I0225 14:11:29.890526 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8ngh\" (UniqueName: \"kubernetes.io/projected/ccaa1a41-0180-48a5-b835-dfcbe76061fb-kube-api-access-x8ngh\") on node \"crc\" DevicePath \"\"" Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.041439 4815 generic.go:334] "Generic (PLEG): container finished" podID="ccaa1a41-0180-48a5-b835-dfcbe76061fb" containerID="5d15df0c5f3622a0b870057ae6af738083bfd3c2d9f20f98143d99152a52351f" exitCode=0 Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.041501 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jdmp9" event={"ID":"ccaa1a41-0180-48a5-b835-dfcbe76061fb","Type":"ContainerDied","Data":"5d15df0c5f3622a0b870057ae6af738083bfd3c2d9f20f98143d99152a52351f"} Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.041599 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jdmp9" event={"ID":"ccaa1a41-0180-48a5-b835-dfcbe76061fb","Type":"ContainerDied","Data":"3b6e70807dea6a024b550b2b5ea3679fd52ed8338d3b979892423ed887e012e0"} Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.041650 4815 scope.go:117] "RemoveContainer" containerID="5d15df0c5f3622a0b870057ae6af738083bfd3c2d9f20f98143d99152a52351f" Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.041663 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jdmp9" Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.073614 4815 scope.go:117] "RemoveContainer" containerID="0a1c3c13699b75ac9a14a38ecb32e7674c4820d009a5ef747c474a80986afd8d" Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.103126 4815 scope.go:117] "RemoveContainer" containerID="39ae088bc8ae38d3a886d4515eea7560ac9eadc2eb080b93ac78b7f135bcbdba" Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.151349 4815 scope.go:117] "RemoveContainer" containerID="5d15df0c5f3622a0b870057ae6af738083bfd3c2d9f20f98143d99152a52351f" Feb 25 14:11:30 crc kubenswrapper[4815]: E0225 14:11:30.152040 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d15df0c5f3622a0b870057ae6af738083bfd3c2d9f20f98143d99152a52351f\": container with ID starting with 5d15df0c5f3622a0b870057ae6af738083bfd3c2d9f20f98143d99152a52351f not found: ID does not exist" containerID="5d15df0c5f3622a0b870057ae6af738083bfd3c2d9f20f98143d99152a52351f" Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.152110 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d15df0c5f3622a0b870057ae6af738083bfd3c2d9f20f98143d99152a52351f"} err="failed to get container status \"5d15df0c5f3622a0b870057ae6af738083bfd3c2d9f20f98143d99152a52351f\": rpc error: code = NotFound desc = could not find container \"5d15df0c5f3622a0b870057ae6af738083bfd3c2d9f20f98143d99152a52351f\": container with ID starting with 5d15df0c5f3622a0b870057ae6af738083bfd3c2d9f20f98143d99152a52351f not found: ID does not exist" Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.152153 4815 scope.go:117] "RemoveContainer" containerID="0a1c3c13699b75ac9a14a38ecb32e7674c4820d009a5ef747c474a80986afd8d" Feb 25 14:11:30 crc kubenswrapper[4815]: E0225 14:11:30.152880 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a1c3c13699b75ac9a14a38ecb32e7674c4820d009a5ef747c474a80986afd8d\": container with ID starting with 0a1c3c13699b75ac9a14a38ecb32e7674c4820d009a5ef747c474a80986afd8d not found: ID does not exist" containerID="0a1c3c13699b75ac9a14a38ecb32e7674c4820d009a5ef747c474a80986afd8d" Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.152940 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a1c3c13699b75ac9a14a38ecb32e7674c4820d009a5ef747c474a80986afd8d"} err="failed to get container status \"0a1c3c13699b75ac9a14a38ecb32e7674c4820d009a5ef747c474a80986afd8d\": rpc error: code = NotFound desc = could not find container \"0a1c3c13699b75ac9a14a38ecb32e7674c4820d009a5ef747c474a80986afd8d\": container with ID starting with 0a1c3c13699b75ac9a14a38ecb32e7674c4820d009a5ef747c474a80986afd8d not found: ID does not exist" Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.152978 4815 scope.go:117] "RemoveContainer" containerID="39ae088bc8ae38d3a886d4515eea7560ac9eadc2eb080b93ac78b7f135bcbdba" Feb 25 14:11:30 crc kubenswrapper[4815]: E0225 14:11:30.153362 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39ae088bc8ae38d3a886d4515eea7560ac9eadc2eb080b93ac78b7f135bcbdba\": container with ID starting with 39ae088bc8ae38d3a886d4515eea7560ac9eadc2eb080b93ac78b7f135bcbdba not found: ID does not exist" containerID="39ae088bc8ae38d3a886d4515eea7560ac9eadc2eb080b93ac78b7f135bcbdba" Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.153432 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39ae088bc8ae38d3a886d4515eea7560ac9eadc2eb080b93ac78b7f135bcbdba"} err="failed to get container status \"39ae088bc8ae38d3a886d4515eea7560ac9eadc2eb080b93ac78b7f135bcbdba\": rpc error: code = NotFound desc = could not find container \"39ae088bc8ae38d3a886d4515eea7560ac9eadc2eb080b93ac78b7f135bcbdba\": container with ID starting with 39ae088bc8ae38d3a886d4515eea7560ac9eadc2eb080b93ac78b7f135bcbdba not found: ID does not exist" Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.239324 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccaa1a41-0180-48a5-b835-dfcbe76061fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ccaa1a41-0180-48a5-b835-dfcbe76061fb" (UID: "ccaa1a41-0180-48a5-b835-dfcbe76061fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.298697 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccaa1a41-0180-48a5-b835-dfcbe76061fb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.386000 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jdmp9"] Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.393636 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jdmp9"] Feb 25 14:11:30 crc kubenswrapper[4815]: I0225 14:11:30.950071 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccaa1a41-0180-48a5-b835-dfcbe76061fb" path="/var/lib/kubelet/pods/ccaa1a41-0180-48a5-b835-dfcbe76061fb/volumes" Feb 25 14:11:42 crc kubenswrapper[4815]: I0225 14:11:42.943899 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:11:42 crc kubenswrapper[4815]: E0225 14:11:42.944720 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:11:53 crc kubenswrapper[4815]: I0225 14:11:53.935767 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:11:53 crc kubenswrapper[4815]: E0225 14:11:53.936587 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:11:59 crc kubenswrapper[4815]: I0225 14:11:59.457744 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m2pk9"] Feb 25 14:11:59 crc kubenswrapper[4815]: E0225 14:11:59.458694 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccaa1a41-0180-48a5-b835-dfcbe76061fb" containerName="registry-server" Feb 25 14:11:59 crc kubenswrapper[4815]: I0225 14:11:59.458706 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccaa1a41-0180-48a5-b835-dfcbe76061fb" containerName="registry-server" Feb 25 14:11:59 crc kubenswrapper[4815]: E0225 14:11:59.458744 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccaa1a41-0180-48a5-b835-dfcbe76061fb" containerName="extract-utilities" Feb 25 14:11:59 crc kubenswrapper[4815]: I0225 14:11:59.458752 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccaa1a41-0180-48a5-b835-dfcbe76061fb" containerName="extract-utilities" Feb 25 14:11:59 crc kubenswrapper[4815]: E0225 14:11:59.458767 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccaa1a41-0180-48a5-b835-dfcbe76061fb" containerName="extract-content" Feb 25 14:11:59 crc kubenswrapper[4815]: I0225 14:11:59.458773 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccaa1a41-0180-48a5-b835-dfcbe76061fb" containerName="extract-content" Feb 25 14:11:59 crc kubenswrapper[4815]: I0225 14:11:59.458936 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccaa1a41-0180-48a5-b835-dfcbe76061fb" containerName="registry-server" Feb 25 14:11:59 crc kubenswrapper[4815]: I0225 14:11:59.467665 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m2pk9" Feb 25 14:11:59 crc kubenswrapper[4815]: I0225 14:11:59.470877 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m2pk9"] Feb 25 14:11:59 crc kubenswrapper[4815]: I0225 14:11:59.585955 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-utilities\") pod \"community-operators-m2pk9\" (UID: \"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93\") " pod="openshift-marketplace/community-operators-m2pk9" Feb 25 14:11:59 crc kubenswrapper[4815]: I0225 14:11:59.586052 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-catalog-content\") pod \"community-operators-m2pk9\" (UID: \"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93\") " pod="openshift-marketplace/community-operators-m2pk9" Feb 25 14:11:59 crc kubenswrapper[4815]: I0225 14:11:59.586118 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msw59\" (UniqueName: \"kubernetes.io/projected/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-kube-api-access-msw59\") pod \"community-operators-m2pk9\" (UID: \"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93\") " pod="openshift-marketplace/community-operators-m2pk9" Feb 25 14:11:59 crc kubenswrapper[4815]: I0225 14:11:59.687815 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-utilities\") pod \"community-operators-m2pk9\" (UID: \"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93\") " pod="openshift-marketplace/community-operators-m2pk9" Feb 25 14:11:59 crc kubenswrapper[4815]: I0225 14:11:59.687913 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-catalog-content\") pod \"community-operators-m2pk9\" (UID: \"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93\") " pod="openshift-marketplace/community-operators-m2pk9" Feb 25 14:11:59 crc kubenswrapper[4815]: I0225 14:11:59.687966 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msw59\" (UniqueName: \"kubernetes.io/projected/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-kube-api-access-msw59\") pod \"community-operators-m2pk9\" (UID: \"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93\") " pod="openshift-marketplace/community-operators-m2pk9" Feb 25 14:11:59 crc kubenswrapper[4815]: I0225 14:11:59.688456 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-utilities\") pod \"community-operators-m2pk9\" (UID: \"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93\") " pod="openshift-marketplace/community-operators-m2pk9" Feb 25 14:11:59 crc kubenswrapper[4815]: I0225 14:11:59.688500 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-catalog-content\") pod \"community-operators-m2pk9\" (UID: \"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93\") " pod="openshift-marketplace/community-operators-m2pk9" Feb 25 14:11:59 crc kubenswrapper[4815]: I0225 14:11:59.716461 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msw59\" (UniqueName: \"kubernetes.io/projected/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-kube-api-access-msw59\") pod \"community-operators-m2pk9\" (UID: \"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93\") " pod="openshift-marketplace/community-operators-m2pk9" Feb 25 14:11:59 crc kubenswrapper[4815]: I0225 14:11:59.796907 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m2pk9" Feb 25 14:12:00 crc kubenswrapper[4815]: I0225 14:12:00.144902 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533812-52hp2"] Feb 25 14:12:00 crc kubenswrapper[4815]: I0225 14:12:00.147224 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533812-52hp2" Feb 25 14:12:00 crc kubenswrapper[4815]: I0225 14:12:00.153197 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 14:12:00 crc kubenswrapper[4815]: I0225 14:12:00.153426 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 14:12:00 crc kubenswrapper[4815]: I0225 14:12:00.153464 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 14:12:00 crc kubenswrapper[4815]: I0225 14:12:00.171134 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533812-52hp2"] Feb 25 14:12:00 crc kubenswrapper[4815]: I0225 14:12:00.299186 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m26nf\" (UniqueName: \"kubernetes.io/projected/78613b37-9fbb-4ef7-8c1c-660da24efdbe-kube-api-access-m26nf\") pod \"auto-csr-approver-29533812-52hp2\" (UID: \"78613b37-9fbb-4ef7-8c1c-660da24efdbe\") " pod="openshift-infra/auto-csr-approver-29533812-52hp2" Feb 25 14:12:00 crc kubenswrapper[4815]: I0225 14:12:00.394140 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m2pk9"] Feb 25 14:12:00 crc kubenswrapper[4815]: I0225 14:12:00.402328 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m26nf\" (UniqueName: \"kubernetes.io/projected/78613b37-9fbb-4ef7-8c1c-660da24efdbe-kube-api-access-m26nf\") pod \"auto-csr-approver-29533812-52hp2\" (UID: \"78613b37-9fbb-4ef7-8c1c-660da24efdbe\") " pod="openshift-infra/auto-csr-approver-29533812-52hp2" Feb 25 14:12:00 crc kubenswrapper[4815]: I0225 14:12:00.428790 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m26nf\" (UniqueName: \"kubernetes.io/projected/78613b37-9fbb-4ef7-8c1c-660da24efdbe-kube-api-access-m26nf\") pod \"auto-csr-approver-29533812-52hp2\" (UID: \"78613b37-9fbb-4ef7-8c1c-660da24efdbe\") " pod="openshift-infra/auto-csr-approver-29533812-52hp2" Feb 25 14:12:00 crc kubenswrapper[4815]: I0225 14:12:00.478623 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533812-52hp2" Feb 25 14:12:00 crc kubenswrapper[4815]: I0225 14:12:00.969475 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533812-52hp2"] Feb 25 14:12:01 crc kubenswrapper[4815]: I0225 14:12:01.407194 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533812-52hp2" event={"ID":"78613b37-9fbb-4ef7-8c1c-660da24efdbe","Type":"ContainerStarted","Data":"c9a5b6461f4d25775a0f6199c542a1e95e5cac8b0d131123971c9978af32357d"} Feb 25 14:12:01 crc kubenswrapper[4815]: I0225 14:12:01.412337 4815 generic.go:334] "Generic (PLEG): container finished" podID="9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93" containerID="313215b641ccaeb6e08a7bd5992172dba1a7045362d975d57f30f2bcf2bf155c" exitCode=0 Feb 25 14:12:01 crc kubenswrapper[4815]: I0225 14:12:01.412385 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2pk9" event={"ID":"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93","Type":"ContainerDied","Data":"313215b641ccaeb6e08a7bd5992172dba1a7045362d975d57f30f2bcf2bf155c"} Feb 25 14:12:01 crc kubenswrapper[4815]: I0225 14:12:01.412412 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2pk9" event={"ID":"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93","Type":"ContainerStarted","Data":"34d0018c2398b3e52e1718b7327de971290894631ed38aac6c20340cb5e3a750"} Feb 25 14:12:02 crc kubenswrapper[4815]: I0225 14:12:02.422669 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2pk9" event={"ID":"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93","Type":"ContainerStarted","Data":"2cd9f2322d425d8056d0650130a2f593f176837b62b73c116edf34bd80e6be56"} Feb 25 14:12:02 crc kubenswrapper[4815]: I0225 14:12:02.424048 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533812-52hp2" event={"ID":"78613b37-9fbb-4ef7-8c1c-660da24efdbe","Type":"ContainerStarted","Data":"9eff1e11700a748ecb9c048ad75b40ddead75959e9d30123375712977a42f6b3"} Feb 25 14:12:02 crc kubenswrapper[4815]: I0225 14:12:02.460406 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533812-52hp2" podStartSLOduration=1.5167270849999999 podStartE2EDuration="2.460389819s" podCreationTimestamp="2026-02-25 14:12:00 +0000 UTC" firstStartedPulling="2026-02-25 14:12:00.951471315 +0000 UTC m=+3078.752569369" lastFinishedPulling="2026-02-25 14:12:01.895134049 +0000 UTC m=+3079.696232103" observedRunningTime="2026-02-25 14:12:02.452843623 +0000 UTC m=+3080.253941677" watchObservedRunningTime="2026-02-25 14:12:02.460389819 +0000 UTC m=+3080.261487873" Feb 25 14:12:03 crc kubenswrapper[4815]: I0225 14:12:03.433841 4815 generic.go:334] "Generic (PLEG): container finished" podID="78613b37-9fbb-4ef7-8c1c-660da24efdbe" containerID="9eff1e11700a748ecb9c048ad75b40ddead75959e9d30123375712977a42f6b3" exitCode=0 Feb 25 14:12:03 crc kubenswrapper[4815]: I0225 14:12:03.433907 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533812-52hp2" event={"ID":"78613b37-9fbb-4ef7-8c1c-660da24efdbe","Type":"ContainerDied","Data":"9eff1e11700a748ecb9c048ad75b40ddead75959e9d30123375712977a42f6b3"} Feb 25 14:12:04 crc kubenswrapper[4815]: I0225 14:12:04.444039 4815 generic.go:334] "Generic (PLEG): container finished" podID="9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93" containerID="2cd9f2322d425d8056d0650130a2f593f176837b62b73c116edf34bd80e6be56" exitCode=0 Feb 25 14:12:04 crc kubenswrapper[4815]: I0225 14:12:04.444213 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2pk9" event={"ID":"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93","Type":"ContainerDied","Data":"2cd9f2322d425d8056d0650130a2f593f176837b62b73c116edf34bd80e6be56"} Feb 25 14:12:04 crc kubenswrapper[4815]: I0225 14:12:04.817671 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533812-52hp2" Feb 25 14:12:04 crc kubenswrapper[4815]: I0225 14:12:04.990949 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m26nf\" (UniqueName: \"kubernetes.io/projected/78613b37-9fbb-4ef7-8c1c-660da24efdbe-kube-api-access-m26nf\") pod \"78613b37-9fbb-4ef7-8c1c-660da24efdbe\" (UID: \"78613b37-9fbb-4ef7-8c1c-660da24efdbe\") " Feb 25 14:12:05 crc kubenswrapper[4815]: I0225 14:12:05.000924 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78613b37-9fbb-4ef7-8c1c-660da24efdbe-kube-api-access-m26nf" (OuterVolumeSpecName: "kube-api-access-m26nf") pod "78613b37-9fbb-4ef7-8c1c-660da24efdbe" (UID: "78613b37-9fbb-4ef7-8c1c-660da24efdbe"). InnerVolumeSpecName "kube-api-access-m26nf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:12:05 crc kubenswrapper[4815]: I0225 14:12:05.093480 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m26nf\" (UniqueName: \"kubernetes.io/projected/78613b37-9fbb-4ef7-8c1c-660da24efdbe-kube-api-access-m26nf\") on node \"crc\" DevicePath \"\"" Feb 25 14:12:05 crc kubenswrapper[4815]: I0225 14:12:05.461850 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533812-52hp2" event={"ID":"78613b37-9fbb-4ef7-8c1c-660da24efdbe","Type":"ContainerDied","Data":"c9a5b6461f4d25775a0f6199c542a1e95e5cac8b0d131123971c9978af32357d"} Feb 25 14:12:05 crc kubenswrapper[4815]: I0225 14:12:05.462127 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9a5b6461f4d25775a0f6199c542a1e95e5cac8b0d131123971c9978af32357d" Feb 25 14:12:05 crc kubenswrapper[4815]: I0225 14:12:05.462192 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533812-52hp2" Feb 25 14:12:05 crc kubenswrapper[4815]: I0225 14:12:05.465491 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2pk9" event={"ID":"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93","Type":"ContainerStarted","Data":"47452f125efb4822dd7caa5dc465798b1f06729ee1af480098298034e7157753"} Feb 25 14:12:05 crc kubenswrapper[4815]: I0225 14:12:05.509568 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m2pk9" podStartSLOduration=2.99534964 podStartE2EDuration="6.509546651s" podCreationTimestamp="2026-02-25 14:11:59 +0000 UTC" firstStartedPulling="2026-02-25 14:12:01.416150612 +0000 UTC m=+3079.217248666" lastFinishedPulling="2026-02-25 14:12:04.930347623 +0000 UTC m=+3082.731445677" observedRunningTime="2026-02-25 14:12:05.486742296 +0000 UTC m=+3083.287840360" watchObservedRunningTime="2026-02-25 14:12:05.509546651 +0000 UTC m=+3083.310644715" Feb 25 14:12:05 crc kubenswrapper[4815]: I0225 14:12:05.542563 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533806-l2585"] Feb 25 14:12:05 crc kubenswrapper[4815]: I0225 14:12:05.550194 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533806-l2585"] Feb 25 14:12:05 crc kubenswrapper[4815]: I0225 14:12:05.935306 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:12:05 crc kubenswrapper[4815]: E0225 14:12:05.935735 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:12:06 crc kubenswrapper[4815]: I0225 14:12:06.948701 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6269a886-839b-41f6-9c68-0759235f4ffc" path="/var/lib/kubelet/pods/6269a886-839b-41f6-9c68-0759235f4ffc/volumes" Feb 25 14:12:09 crc kubenswrapper[4815]: I0225 14:12:09.798073 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m2pk9" Feb 25 14:12:09 crc kubenswrapper[4815]: I0225 14:12:09.798401 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m2pk9" Feb 25 14:12:09 crc kubenswrapper[4815]: I0225 14:12:09.860902 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m2pk9" Feb 25 14:12:10 crc kubenswrapper[4815]: I0225 14:12:10.571914 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m2pk9" Feb 25 14:12:10 crc kubenswrapper[4815]: I0225 14:12:10.619350 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m2pk9"] Feb 25 14:12:12 crc kubenswrapper[4815]: I0225 14:12:12.529896 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m2pk9" podUID="9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93" containerName="registry-server" containerID="cri-o://47452f125efb4822dd7caa5dc465798b1f06729ee1af480098298034e7157753" gracePeriod=2 Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.084192 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m2pk9" Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.164543 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msw59\" (UniqueName: \"kubernetes.io/projected/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-kube-api-access-msw59\") pod \"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93\" (UID: \"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93\") " Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.165197 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-utilities\") pod \"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93\" (UID: \"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93\") " Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.165334 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-catalog-content\") pod \"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93\" (UID: \"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93\") " Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.166613 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-utilities" (OuterVolumeSpecName: "utilities") pod "9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93" (UID: "9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.175057 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-kube-api-access-msw59" (OuterVolumeSpecName: "kube-api-access-msw59") pod "9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93" (UID: "9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93"). InnerVolumeSpecName "kube-api-access-msw59". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.229478 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93" (UID: "9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.267746 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.267782 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.267795 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msw59\" (UniqueName: \"kubernetes.io/projected/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93-kube-api-access-msw59\") on node \"crc\" DevicePath \"\"" Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.543583 4815 generic.go:334] "Generic (PLEG): container finished" podID="9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93" containerID="47452f125efb4822dd7caa5dc465798b1f06729ee1af480098298034e7157753" exitCode=0 Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.543648 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m2pk9" Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.543653 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2pk9" event={"ID":"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93","Type":"ContainerDied","Data":"47452f125efb4822dd7caa5dc465798b1f06729ee1af480098298034e7157753"} Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.545085 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2pk9" event={"ID":"9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93","Type":"ContainerDied","Data":"34d0018c2398b3e52e1718b7327de971290894631ed38aac6c20340cb5e3a750"} Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.545176 4815 scope.go:117] "RemoveContainer" containerID="47452f125efb4822dd7caa5dc465798b1f06729ee1af480098298034e7157753" Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.571389 4815 scope.go:117] "RemoveContainer" containerID="2cd9f2322d425d8056d0650130a2f593f176837b62b73c116edf34bd80e6be56" Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.587282 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m2pk9"] Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.596028 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m2pk9"] Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.611032 4815 scope.go:117] "RemoveContainer" containerID="313215b641ccaeb6e08a7bd5992172dba1a7045362d975d57f30f2bcf2bf155c" Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.638683 4815 scope.go:117] "RemoveContainer" containerID="47452f125efb4822dd7caa5dc465798b1f06729ee1af480098298034e7157753" Feb 25 14:12:13 crc kubenswrapper[4815]: E0225 14:12:13.639248 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47452f125efb4822dd7caa5dc465798b1f06729ee1af480098298034e7157753\": container with ID starting with 47452f125efb4822dd7caa5dc465798b1f06729ee1af480098298034e7157753 not found: ID does not exist" containerID="47452f125efb4822dd7caa5dc465798b1f06729ee1af480098298034e7157753" Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.639303 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47452f125efb4822dd7caa5dc465798b1f06729ee1af480098298034e7157753"} err="failed to get container status \"47452f125efb4822dd7caa5dc465798b1f06729ee1af480098298034e7157753\": rpc error: code = NotFound desc = could not find container \"47452f125efb4822dd7caa5dc465798b1f06729ee1af480098298034e7157753\": container with ID starting with 47452f125efb4822dd7caa5dc465798b1f06729ee1af480098298034e7157753 not found: ID does not exist" Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.639448 4815 scope.go:117] "RemoveContainer" containerID="2cd9f2322d425d8056d0650130a2f593f176837b62b73c116edf34bd80e6be56" Feb 25 14:12:13 crc kubenswrapper[4815]: E0225 14:12:13.639850 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cd9f2322d425d8056d0650130a2f593f176837b62b73c116edf34bd80e6be56\": container with ID starting with 2cd9f2322d425d8056d0650130a2f593f176837b62b73c116edf34bd80e6be56 not found: ID does not exist" containerID="2cd9f2322d425d8056d0650130a2f593f176837b62b73c116edf34bd80e6be56" Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.639881 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cd9f2322d425d8056d0650130a2f593f176837b62b73c116edf34bd80e6be56"} err="failed to get container status \"2cd9f2322d425d8056d0650130a2f593f176837b62b73c116edf34bd80e6be56\": rpc error: code = NotFound desc = could not find container \"2cd9f2322d425d8056d0650130a2f593f176837b62b73c116edf34bd80e6be56\": container with ID starting with 2cd9f2322d425d8056d0650130a2f593f176837b62b73c116edf34bd80e6be56 not found: ID does not exist" Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.639898 4815 scope.go:117] "RemoveContainer" containerID="313215b641ccaeb6e08a7bd5992172dba1a7045362d975d57f30f2bcf2bf155c" Feb 25 14:12:13 crc kubenswrapper[4815]: E0225 14:12:13.640121 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"313215b641ccaeb6e08a7bd5992172dba1a7045362d975d57f30f2bcf2bf155c\": container with ID starting with 313215b641ccaeb6e08a7bd5992172dba1a7045362d975d57f30f2bcf2bf155c not found: ID does not exist" containerID="313215b641ccaeb6e08a7bd5992172dba1a7045362d975d57f30f2bcf2bf155c" Feb 25 14:12:13 crc kubenswrapper[4815]: I0225 14:12:13.640149 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"313215b641ccaeb6e08a7bd5992172dba1a7045362d975d57f30f2bcf2bf155c"} err="failed to get container status \"313215b641ccaeb6e08a7bd5992172dba1a7045362d975d57f30f2bcf2bf155c\": rpc error: code = NotFound desc = could not find container \"313215b641ccaeb6e08a7bd5992172dba1a7045362d975d57f30f2bcf2bf155c\": container with ID starting with 313215b641ccaeb6e08a7bd5992172dba1a7045362d975d57f30f2bcf2bf155c not found: ID does not exist" Feb 25 14:12:14 crc kubenswrapper[4815]: I0225 14:12:14.966994 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93" path="/var/lib/kubelet/pods/9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93/volumes" Feb 25 14:12:17 crc kubenswrapper[4815]: I0225 14:12:17.937005 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:12:17 crc kubenswrapper[4815]: E0225 14:12:17.937781 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:12:30 crc kubenswrapper[4815]: I0225 14:12:30.936683 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:12:30 crc kubenswrapper[4815]: E0225 14:12:30.939667 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:12:42 crc kubenswrapper[4815]: I0225 14:12:42.943016 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:12:42 crc kubenswrapper[4815]: E0225 14:12:42.944365 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:12:49 crc kubenswrapper[4815]: I0225 14:12:49.465141 4815 scope.go:117] "RemoveContainer" containerID="a34586a0d45fed6749a23f18bd02304a1e6eedb25b5892dfa0f8043ba4597263" Feb 25 14:12:57 crc kubenswrapper[4815]: I0225 14:12:57.935446 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:12:57 crc kubenswrapper[4815]: E0225 14:12:57.936169 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:13:08 crc kubenswrapper[4815]: I0225 14:13:08.936421 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:13:08 crc kubenswrapper[4815]: E0225 14:13:08.937308 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:13:20 crc kubenswrapper[4815]: I0225 14:13:20.935957 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:13:20 crc kubenswrapper[4815]: E0225 14:13:20.936839 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:13:31 crc kubenswrapper[4815]: I0225 14:13:31.936790 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:13:31 crc kubenswrapper[4815]: E0225 14:13:31.937863 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:13:43 crc kubenswrapper[4815]: I0225 14:13:43.936400 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:13:43 crc kubenswrapper[4815]: E0225 14:13:43.937321 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:13:54 crc kubenswrapper[4815]: I0225 14:13:54.936262 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:13:54 crc kubenswrapper[4815]: E0225 14:13:54.937014 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:14:00 crc kubenswrapper[4815]: I0225 14:14:00.196860 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533814-mjcm4"] Feb 25 14:14:00 crc kubenswrapper[4815]: E0225 14:14:00.197779 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93" containerName="extract-utilities" Feb 25 14:14:00 crc kubenswrapper[4815]: I0225 14:14:00.197795 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93" containerName="extract-utilities" Feb 25 14:14:00 crc kubenswrapper[4815]: E0225 14:14:00.197818 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93" containerName="extract-content" Feb 25 14:14:00 crc kubenswrapper[4815]: I0225 14:14:00.197826 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93" containerName="extract-content" Feb 25 14:14:00 crc kubenswrapper[4815]: E0225 14:14:00.197835 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93" containerName="registry-server" Feb 25 14:14:00 crc kubenswrapper[4815]: I0225 14:14:00.197842 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93" containerName="registry-server" Feb 25 14:14:00 crc kubenswrapper[4815]: E0225 14:14:00.197857 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78613b37-9fbb-4ef7-8c1c-660da24efdbe" containerName="oc" Feb 25 14:14:00 crc kubenswrapper[4815]: I0225 14:14:00.197864 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="78613b37-9fbb-4ef7-8c1c-660da24efdbe" containerName="oc" Feb 25 14:14:00 crc kubenswrapper[4815]: I0225 14:14:00.198111 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="78613b37-9fbb-4ef7-8c1c-660da24efdbe" containerName="oc" Feb 25 14:14:00 crc kubenswrapper[4815]: I0225 14:14:00.198128 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="9629e2e0-e6a2-4733-9fed-c6f7b4b4dc93" containerName="registry-server" Feb 25 14:14:00 crc kubenswrapper[4815]: I0225 14:14:00.198970 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533814-mjcm4" Feb 25 14:14:00 crc kubenswrapper[4815]: I0225 14:14:00.201836 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 14:14:00 crc kubenswrapper[4815]: I0225 14:14:00.202038 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 14:14:00 crc kubenswrapper[4815]: I0225 14:14:00.202480 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 14:14:00 crc kubenswrapper[4815]: I0225 14:14:00.215055 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533814-mjcm4"] Feb 25 14:14:00 crc kubenswrapper[4815]: I0225 14:14:00.271636 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2rbh\" (UniqueName: \"kubernetes.io/projected/f1e09176-fa9b-4ace-bb60-ee5f54de9571-kube-api-access-j2rbh\") pod \"auto-csr-approver-29533814-mjcm4\" (UID: \"f1e09176-fa9b-4ace-bb60-ee5f54de9571\") " pod="openshift-infra/auto-csr-approver-29533814-mjcm4" Feb 25 14:14:00 crc kubenswrapper[4815]: I0225 14:14:00.373386 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2rbh\" (UniqueName: \"kubernetes.io/projected/f1e09176-fa9b-4ace-bb60-ee5f54de9571-kube-api-access-j2rbh\") pod \"auto-csr-approver-29533814-mjcm4\" (UID: \"f1e09176-fa9b-4ace-bb60-ee5f54de9571\") " pod="openshift-infra/auto-csr-approver-29533814-mjcm4" Feb 25 14:14:00 crc kubenswrapper[4815]: I0225 14:14:00.393793 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2rbh\" (UniqueName: \"kubernetes.io/projected/f1e09176-fa9b-4ace-bb60-ee5f54de9571-kube-api-access-j2rbh\") pod \"auto-csr-approver-29533814-mjcm4\" (UID: \"f1e09176-fa9b-4ace-bb60-ee5f54de9571\") " pod="openshift-infra/auto-csr-approver-29533814-mjcm4" Feb 25 14:14:00 crc kubenswrapper[4815]: I0225 14:14:00.520539 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533814-mjcm4" Feb 25 14:14:00 crc kubenswrapper[4815]: I0225 14:14:00.963471 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533814-mjcm4"] Feb 25 14:14:01 crc kubenswrapper[4815]: I0225 14:14:01.565170 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533814-mjcm4" event={"ID":"f1e09176-fa9b-4ace-bb60-ee5f54de9571","Type":"ContainerStarted","Data":"bb332638773d49d98208adb77aa4341fce907bcd4811db2f0bb3d7a3d1132ae8"} Feb 25 14:14:02 crc kubenswrapper[4815]: I0225 14:14:02.574990 4815 generic.go:334] "Generic (PLEG): container finished" podID="f1e09176-fa9b-4ace-bb60-ee5f54de9571" containerID="a4b9a30738cf19ad63e8a58ea4fd11a35641d4d44704d0f3755d1ca1d18aed46" exitCode=0 Feb 25 14:14:02 crc kubenswrapper[4815]: I0225 14:14:02.575055 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533814-mjcm4" event={"ID":"f1e09176-fa9b-4ace-bb60-ee5f54de9571","Type":"ContainerDied","Data":"a4b9a30738cf19ad63e8a58ea4fd11a35641d4d44704d0f3755d1ca1d18aed46"} Feb 25 14:14:03 crc kubenswrapper[4815]: I0225 14:14:03.982013 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533814-mjcm4" Feb 25 14:14:04 crc kubenswrapper[4815]: I0225 14:14:04.049186 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2rbh\" (UniqueName: \"kubernetes.io/projected/f1e09176-fa9b-4ace-bb60-ee5f54de9571-kube-api-access-j2rbh\") pod \"f1e09176-fa9b-4ace-bb60-ee5f54de9571\" (UID: \"f1e09176-fa9b-4ace-bb60-ee5f54de9571\") " Feb 25 14:14:04 crc kubenswrapper[4815]: I0225 14:14:04.056209 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1e09176-fa9b-4ace-bb60-ee5f54de9571-kube-api-access-j2rbh" (OuterVolumeSpecName: "kube-api-access-j2rbh") pod "f1e09176-fa9b-4ace-bb60-ee5f54de9571" (UID: "f1e09176-fa9b-4ace-bb60-ee5f54de9571"). InnerVolumeSpecName "kube-api-access-j2rbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:14:04 crc kubenswrapper[4815]: I0225 14:14:04.151605 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2rbh\" (UniqueName: \"kubernetes.io/projected/f1e09176-fa9b-4ace-bb60-ee5f54de9571-kube-api-access-j2rbh\") on node \"crc\" DevicePath \"\"" Feb 25 14:14:04 crc kubenswrapper[4815]: I0225 14:14:04.593054 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533814-mjcm4" event={"ID":"f1e09176-fa9b-4ace-bb60-ee5f54de9571","Type":"ContainerDied","Data":"bb332638773d49d98208adb77aa4341fce907bcd4811db2f0bb3d7a3d1132ae8"} Feb 25 14:14:04 crc kubenswrapper[4815]: I0225 14:14:04.593092 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb332638773d49d98208adb77aa4341fce907bcd4811db2f0bb3d7a3d1132ae8" Feb 25 14:14:04 crc kubenswrapper[4815]: I0225 14:14:04.593111 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533814-mjcm4" Feb 25 14:14:05 crc kubenswrapper[4815]: I0225 14:14:05.083274 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533808-nlxhc"] Feb 25 14:14:05 crc kubenswrapper[4815]: I0225 14:14:05.101078 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533808-nlxhc"] Feb 25 14:14:06 crc kubenswrapper[4815]: I0225 14:14:06.936269 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:14:06 crc kubenswrapper[4815]: E0225 14:14:06.936933 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:14:06 crc kubenswrapper[4815]: I0225 14:14:06.957131 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90489231-445c-4533-833e-3c0116ff824e" path="/var/lib/kubelet/pods/90489231-445c-4533-833e-3c0116ff824e/volumes" Feb 25 14:14:21 crc kubenswrapper[4815]: I0225 14:14:21.936158 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:14:22 crc kubenswrapper[4815]: I0225 14:14:22.756934 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerStarted","Data":"35ad0ed6ac05ca7660eb380c1f3eb06169b51f834d4f9b66e462803955bc5ff9"} Feb 25 14:14:41 crc kubenswrapper[4815]: I0225 14:14:41.930463 4815 generic.go:334] "Generic (PLEG): container finished" podID="d23e8f04-361c-4227-a5c9-bfcd4801f771" containerID="3cf1f2d98cbdd054697aa239331c4744120fbfbd01368da1d5dbd4c6ce8bd28f" exitCode=1 Feb 25 14:14:41 crc kubenswrapper[4815]: I0225 14:14:41.930580 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d23e8f04-361c-4227-a5c9-bfcd4801f771","Type":"ContainerDied","Data":"3cf1f2d98cbdd054697aa239331c4744120fbfbd01368da1d5dbd4c6ce8bd28f"} Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.411954 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.543555 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"d23e8f04-361c-4227-a5c9-bfcd4801f771\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.543611 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqkvk\" (UniqueName: \"kubernetes.io/projected/d23e8f04-361c-4227-a5c9-bfcd4801f771-kube-api-access-lqkvk\") pod \"d23e8f04-361c-4227-a5c9-bfcd4801f771\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.543641 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-ssh-key\") pod \"d23e8f04-361c-4227-a5c9-bfcd4801f771\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.543661 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d23e8f04-361c-4227-a5c9-bfcd4801f771-openstack-config\") pod \"d23e8f04-361c-4227-a5c9-bfcd4801f771\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.543865 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d23e8f04-361c-4227-a5c9-bfcd4801f771-test-operator-ephemeral-temporary\") pod \"d23e8f04-361c-4227-a5c9-bfcd4801f771\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.543899 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d23e8f04-361c-4227-a5c9-bfcd4801f771-test-operator-ephemeral-workdir\") pod \"d23e8f04-361c-4227-a5c9-bfcd4801f771\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.543917 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-openstack-config-secret\") pod \"d23e8f04-361c-4227-a5c9-bfcd4801f771\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.543947 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d23e8f04-361c-4227-a5c9-bfcd4801f771-config-data\") pod \"d23e8f04-361c-4227-a5c9-bfcd4801f771\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.543965 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-ca-certs\") pod \"d23e8f04-361c-4227-a5c9-bfcd4801f771\" (UID: \"d23e8f04-361c-4227-a5c9-bfcd4801f771\") " Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.545197 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d23e8f04-361c-4227-a5c9-bfcd4801f771-config-data" (OuterVolumeSpecName: "config-data") pod "d23e8f04-361c-4227-a5c9-bfcd4801f771" (UID: "d23e8f04-361c-4227-a5c9-bfcd4801f771"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.545399 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d23e8f04-361c-4227-a5c9-bfcd4801f771-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "d23e8f04-361c-4227-a5c9-bfcd4801f771" (UID: "d23e8f04-361c-4227-a5c9-bfcd4801f771"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.549613 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d23e8f04-361c-4227-a5c9-bfcd4801f771-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "d23e8f04-361c-4227-a5c9-bfcd4801f771" (UID: "d23e8f04-361c-4227-a5c9-bfcd4801f771"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.550233 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "test-operator-logs") pod "d23e8f04-361c-4227-a5c9-bfcd4801f771" (UID: "d23e8f04-361c-4227-a5c9-bfcd4801f771"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.568556 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d23e8f04-361c-4227-a5c9-bfcd4801f771-kube-api-access-lqkvk" (OuterVolumeSpecName: "kube-api-access-lqkvk") pod "d23e8f04-361c-4227-a5c9-bfcd4801f771" (UID: "d23e8f04-361c-4227-a5c9-bfcd4801f771"). InnerVolumeSpecName "kube-api-access-lqkvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.577645 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "d23e8f04-361c-4227-a5c9-bfcd4801f771" (UID: "d23e8f04-361c-4227-a5c9-bfcd4801f771"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.583702 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d23e8f04-361c-4227-a5c9-bfcd4801f771" (UID: "d23e8f04-361c-4227-a5c9-bfcd4801f771"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.585916 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "d23e8f04-361c-4227-a5c9-bfcd4801f771" (UID: "d23e8f04-361c-4227-a5c9-bfcd4801f771"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.597111 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d23e8f04-361c-4227-a5c9-bfcd4801f771-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "d23e8f04-361c-4227-a5c9-bfcd4801f771" (UID: "d23e8f04-361c-4227-a5c9-bfcd4801f771"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.646242 4815 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.646399 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqkvk\" (UniqueName: \"kubernetes.io/projected/d23e8f04-361c-4227-a5c9-bfcd4801f771-kube-api-access-lqkvk\") on node \"crc\" DevicePath \"\"" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.646471 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-ssh-key\") on node \"crc\" DevicePath \"\"" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.646610 4815 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d23e8f04-361c-4227-a5c9-bfcd4801f771-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.646669 4815 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d23e8f04-361c-4227-a5c9-bfcd4801f771-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.646724 4815 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.646807 4815 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d23e8f04-361c-4227-a5c9-bfcd4801f771-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.646872 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d23e8f04-361c-4227-a5c9-bfcd4801f771-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.646923 4815 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d23e8f04-361c-4227-a5c9-bfcd4801f771-ca-certs\") on node \"crc\" DevicePath \"\"" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.665455 4815 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.749199 4815 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.967072 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d23e8f04-361c-4227-a5c9-bfcd4801f771","Type":"ContainerDied","Data":"84b2df7e536fa12971c21c2080bccea901c67f26be53d99e35ec961c4e9d9a4b"} Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.967135 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84b2df7e536fa12971c21c2080bccea901c67f26be53d99e35ec961c4e9d9a4b" Feb 25 14:14:43 crc kubenswrapper[4815]: I0225 14:14:43.967220 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 25 14:14:49 crc kubenswrapper[4815]: I0225 14:14:49.607538 4815 scope.go:117] "RemoveContainer" containerID="3bc65c3a29af0bc97a6740353fa892252c3148b0388e77701f0b4bb5beca3925" Feb 25 14:14:53 crc kubenswrapper[4815]: I0225 14:14:53.792805 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 25 14:14:53 crc kubenswrapper[4815]: E0225 14:14:53.794127 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d23e8f04-361c-4227-a5c9-bfcd4801f771" containerName="tempest-tests-tempest-tests-runner" Feb 25 14:14:53 crc kubenswrapper[4815]: I0225 14:14:53.794153 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d23e8f04-361c-4227-a5c9-bfcd4801f771" containerName="tempest-tests-tempest-tests-runner" Feb 25 14:14:53 crc kubenswrapper[4815]: E0225 14:14:53.794183 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e09176-fa9b-4ace-bb60-ee5f54de9571" containerName="oc" Feb 25 14:14:53 crc kubenswrapper[4815]: I0225 14:14:53.794196 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e09176-fa9b-4ace-bb60-ee5f54de9571" containerName="oc" Feb 25 14:14:53 crc kubenswrapper[4815]: I0225 14:14:53.794650 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e09176-fa9b-4ace-bb60-ee5f54de9571" containerName="oc" Feb 25 14:14:53 crc kubenswrapper[4815]: I0225 14:14:53.794682 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="d23e8f04-361c-4227-a5c9-bfcd4801f771" containerName="tempest-tests-tempest-tests-runner" Feb 25 14:14:53 crc kubenswrapper[4815]: I0225 14:14:53.795784 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 14:14:53 crc kubenswrapper[4815]: I0225 14:14:53.799586 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-sqqf7" Feb 25 14:14:53 crc kubenswrapper[4815]: I0225 14:14:53.806098 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 25 14:14:53 crc kubenswrapper[4815]: I0225 14:14:53.862692 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"91988324-9ed0-4293-82e5-738a92f680e1\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 14:14:53 crc kubenswrapper[4815]: I0225 14:14:53.863192 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjkzm\" (UniqueName: \"kubernetes.io/projected/91988324-9ed0-4293-82e5-738a92f680e1-kube-api-access-bjkzm\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"91988324-9ed0-4293-82e5-738a92f680e1\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 14:14:53 crc kubenswrapper[4815]: I0225 14:14:53.965297 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"91988324-9ed0-4293-82e5-738a92f680e1\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 14:14:53 crc kubenswrapper[4815]: I0225 14:14:53.965458 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjkzm\" (UniqueName: \"kubernetes.io/projected/91988324-9ed0-4293-82e5-738a92f680e1-kube-api-access-bjkzm\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"91988324-9ed0-4293-82e5-738a92f680e1\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 14:14:53 crc kubenswrapper[4815]: I0225 14:14:53.966758 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"91988324-9ed0-4293-82e5-738a92f680e1\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 14:14:53 crc kubenswrapper[4815]: I0225 14:14:53.994289 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjkzm\" (UniqueName: \"kubernetes.io/projected/91988324-9ed0-4293-82e5-738a92f680e1-kube-api-access-bjkzm\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"91988324-9ed0-4293-82e5-738a92f680e1\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 14:14:54 crc kubenswrapper[4815]: I0225 14:14:54.024955 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"91988324-9ed0-4293-82e5-738a92f680e1\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 14:14:54 crc kubenswrapper[4815]: I0225 14:14:54.133850 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 14:14:54 crc kubenswrapper[4815]: I0225 14:14:54.629813 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 25 14:14:55 crc kubenswrapper[4815]: I0225 14:14:55.092338 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"91988324-9ed0-4293-82e5-738a92f680e1","Type":"ContainerStarted","Data":"0a1a28867f8843441affab4a70ac175e565511b6f04c18bffb9beacd3ff29b76"} Feb 25 14:14:56 crc kubenswrapper[4815]: I0225 14:14:56.107569 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"91988324-9ed0-4293-82e5-738a92f680e1","Type":"ContainerStarted","Data":"529b00a3e2682b297562e77eb88457780e1fbfa6df5ac332e704f9d9a5bbea5d"} Feb 25 14:14:56 crc kubenswrapper[4815]: I0225 14:14:56.132207 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.298521413 podStartE2EDuration="3.132182658s" podCreationTimestamp="2026-02-25 14:14:53 +0000 UTC" firstStartedPulling="2026-02-25 14:14:54.634094063 +0000 UTC m=+3252.435192127" lastFinishedPulling="2026-02-25 14:14:55.467755278 +0000 UTC m=+3253.268853372" observedRunningTime="2026-02-25 14:14:56.126467779 +0000 UTC m=+3253.927565863" watchObservedRunningTime="2026-02-25 14:14:56.132182658 +0000 UTC m=+3253.933280722" Feb 25 14:15:00 crc kubenswrapper[4815]: I0225 14:15:00.160783 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7"] Feb 25 14:15:00 crc kubenswrapper[4815]: I0225 14:15:00.162766 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" Feb 25 14:15:00 crc kubenswrapper[4815]: I0225 14:15:00.164890 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 14:15:00 crc kubenswrapper[4815]: I0225 14:15:00.165557 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 14:15:00 crc kubenswrapper[4815]: I0225 14:15:00.173111 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7"] Feb 25 14:15:00 crc kubenswrapper[4815]: I0225 14:15:00.215318 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5wx9\" (UniqueName: \"kubernetes.io/projected/0cc0d985-5b1d-4756-ac53-be6138bb46ba-kube-api-access-b5wx9\") pod \"collect-profiles-29533815-gspp7\" (UID: \"0cc0d985-5b1d-4756-ac53-be6138bb46ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" Feb 25 14:15:00 crc kubenswrapper[4815]: I0225 14:15:00.215735 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0cc0d985-5b1d-4756-ac53-be6138bb46ba-secret-volume\") pod \"collect-profiles-29533815-gspp7\" (UID: \"0cc0d985-5b1d-4756-ac53-be6138bb46ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" Feb 25 14:15:00 crc kubenswrapper[4815]: I0225 14:15:00.215922 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0cc0d985-5b1d-4756-ac53-be6138bb46ba-config-volume\") pod \"collect-profiles-29533815-gspp7\" (UID: \"0cc0d985-5b1d-4756-ac53-be6138bb46ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" Feb 25 14:15:00 crc kubenswrapper[4815]: I0225 14:15:00.317799 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5wx9\" (UniqueName: \"kubernetes.io/projected/0cc0d985-5b1d-4756-ac53-be6138bb46ba-kube-api-access-b5wx9\") pod \"collect-profiles-29533815-gspp7\" (UID: \"0cc0d985-5b1d-4756-ac53-be6138bb46ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" Feb 25 14:15:00 crc kubenswrapper[4815]: I0225 14:15:00.318201 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0cc0d985-5b1d-4756-ac53-be6138bb46ba-secret-volume\") pod \"collect-profiles-29533815-gspp7\" (UID: \"0cc0d985-5b1d-4756-ac53-be6138bb46ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" Feb 25 14:15:00 crc kubenswrapper[4815]: I0225 14:15:00.319124 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0cc0d985-5b1d-4756-ac53-be6138bb46ba-config-volume\") pod \"collect-profiles-29533815-gspp7\" (UID: \"0cc0d985-5b1d-4756-ac53-be6138bb46ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" Feb 25 14:15:00 crc kubenswrapper[4815]: I0225 14:15:00.320128 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0cc0d985-5b1d-4756-ac53-be6138bb46ba-config-volume\") pod \"collect-profiles-29533815-gspp7\" (UID: \"0cc0d985-5b1d-4756-ac53-be6138bb46ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" Feb 25 14:15:00 crc kubenswrapper[4815]: I0225 14:15:00.325336 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0cc0d985-5b1d-4756-ac53-be6138bb46ba-secret-volume\") pod \"collect-profiles-29533815-gspp7\" (UID: \"0cc0d985-5b1d-4756-ac53-be6138bb46ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" Feb 25 14:15:00 crc kubenswrapper[4815]: I0225 14:15:00.340336 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5wx9\" (UniqueName: \"kubernetes.io/projected/0cc0d985-5b1d-4756-ac53-be6138bb46ba-kube-api-access-b5wx9\") pod \"collect-profiles-29533815-gspp7\" (UID: \"0cc0d985-5b1d-4756-ac53-be6138bb46ba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" Feb 25 14:15:00 crc kubenswrapper[4815]: I0225 14:15:00.489350 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" Feb 25 14:15:00 crc kubenswrapper[4815]: I0225 14:15:00.920571 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7"] Feb 25 14:15:00 crc kubenswrapper[4815]: W0225 14:15:00.922268 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cc0d985_5b1d_4756_ac53_be6138bb46ba.slice/crio-980d866d893e5b4ebe346901e381601ebd3f5791fa57c2b6835a2c3de6b6a371 WatchSource:0}: Error finding container 980d866d893e5b4ebe346901e381601ebd3f5791fa57c2b6835a2c3de6b6a371: Status 404 returned error can't find the container with id 980d866d893e5b4ebe346901e381601ebd3f5791fa57c2b6835a2c3de6b6a371 Feb 25 14:15:01 crc kubenswrapper[4815]: I0225 14:15:01.158128 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" event={"ID":"0cc0d985-5b1d-4756-ac53-be6138bb46ba","Type":"ContainerStarted","Data":"208bf39f101626f8941c61f09c9d8e49b0df2cb082c5c0ef1ac056c0a92b41af"} Feb 25 14:15:01 crc kubenswrapper[4815]: I0225 14:15:01.158186 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" event={"ID":"0cc0d985-5b1d-4756-ac53-be6138bb46ba","Type":"ContainerStarted","Data":"980d866d893e5b4ebe346901e381601ebd3f5791fa57c2b6835a2c3de6b6a371"} Feb 25 14:15:01 crc kubenswrapper[4815]: I0225 14:15:01.188580 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" podStartSLOduration=1.188558146 podStartE2EDuration="1.188558146s" podCreationTimestamp="2026-02-25 14:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 14:15:01.173637158 +0000 UTC m=+3258.974735212" watchObservedRunningTime="2026-02-25 14:15:01.188558146 +0000 UTC m=+3258.989656210" Feb 25 14:15:02 crc kubenswrapper[4815]: I0225 14:15:02.168298 4815 generic.go:334] "Generic (PLEG): container finished" podID="0cc0d985-5b1d-4756-ac53-be6138bb46ba" containerID="208bf39f101626f8941c61f09c9d8e49b0df2cb082c5c0ef1ac056c0a92b41af" exitCode=0 Feb 25 14:15:02 crc kubenswrapper[4815]: I0225 14:15:02.168351 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" event={"ID":"0cc0d985-5b1d-4756-ac53-be6138bb46ba","Type":"ContainerDied","Data":"208bf39f101626f8941c61f09c9d8e49b0df2cb082c5c0ef1ac056c0a92b41af"} Feb 25 14:15:03 crc kubenswrapper[4815]: I0225 14:15:03.464342 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" Feb 25 14:15:03 crc kubenswrapper[4815]: I0225 14:15:03.600415 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5wx9\" (UniqueName: \"kubernetes.io/projected/0cc0d985-5b1d-4756-ac53-be6138bb46ba-kube-api-access-b5wx9\") pod \"0cc0d985-5b1d-4756-ac53-be6138bb46ba\" (UID: \"0cc0d985-5b1d-4756-ac53-be6138bb46ba\") " Feb 25 14:15:03 crc kubenswrapper[4815]: I0225 14:15:03.600761 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0cc0d985-5b1d-4756-ac53-be6138bb46ba-config-volume\") pod \"0cc0d985-5b1d-4756-ac53-be6138bb46ba\" (UID: \"0cc0d985-5b1d-4756-ac53-be6138bb46ba\") " Feb 25 14:15:03 crc kubenswrapper[4815]: I0225 14:15:03.601071 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0cc0d985-5b1d-4756-ac53-be6138bb46ba-secret-volume\") pod \"0cc0d985-5b1d-4756-ac53-be6138bb46ba\" (UID: \"0cc0d985-5b1d-4756-ac53-be6138bb46ba\") " Feb 25 14:15:03 crc kubenswrapper[4815]: I0225 14:15:03.601992 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cc0d985-5b1d-4756-ac53-be6138bb46ba-config-volume" (OuterVolumeSpecName: "config-volume") pod "0cc0d985-5b1d-4756-ac53-be6138bb46ba" (UID: "0cc0d985-5b1d-4756-ac53-be6138bb46ba"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 14:15:03 crc kubenswrapper[4815]: I0225 14:15:03.602311 4815 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0cc0d985-5b1d-4756-ac53-be6138bb46ba-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 14:15:03 crc kubenswrapper[4815]: I0225 14:15:03.608049 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cc0d985-5b1d-4756-ac53-be6138bb46ba-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0cc0d985-5b1d-4756-ac53-be6138bb46ba" (UID: "0cc0d985-5b1d-4756-ac53-be6138bb46ba"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 14:15:03 crc kubenswrapper[4815]: I0225 14:15:03.609310 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cc0d985-5b1d-4756-ac53-be6138bb46ba-kube-api-access-b5wx9" (OuterVolumeSpecName: "kube-api-access-b5wx9") pod "0cc0d985-5b1d-4756-ac53-be6138bb46ba" (UID: "0cc0d985-5b1d-4756-ac53-be6138bb46ba"). InnerVolumeSpecName "kube-api-access-b5wx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:15:03 crc kubenswrapper[4815]: I0225 14:15:03.703968 4815 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0cc0d985-5b1d-4756-ac53-be6138bb46ba-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 14:15:03 crc kubenswrapper[4815]: I0225 14:15:03.703998 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5wx9\" (UniqueName: \"kubernetes.io/projected/0cc0d985-5b1d-4756-ac53-be6138bb46ba-kube-api-access-b5wx9\") on node \"crc\" DevicePath \"\"" Feb 25 14:15:04 crc kubenswrapper[4815]: I0225 14:15:04.190184 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" event={"ID":"0cc0d985-5b1d-4756-ac53-be6138bb46ba","Type":"ContainerDied","Data":"980d866d893e5b4ebe346901e381601ebd3f5791fa57c2b6835a2c3de6b6a371"} Feb 25 14:15:04 crc kubenswrapper[4815]: I0225 14:15:04.190529 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="980d866d893e5b4ebe346901e381601ebd3f5791fa57c2b6835a2c3de6b6a371" Feb 25 14:15:04 crc kubenswrapper[4815]: I0225 14:15:04.190238 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533815-gspp7" Feb 25 14:15:04 crc kubenswrapper[4815]: I0225 14:15:04.257271 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt"] Feb 25 14:15:04 crc kubenswrapper[4815]: I0225 14:15:04.264769 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533770-sjpgt"] Feb 25 14:15:04 crc kubenswrapper[4815]: I0225 14:15:04.957295 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6f07757-816a-4dd2-b804-795dd6a874fd" path="/var/lib/kubelet/pods/b6f07757-816a-4dd2-b804-795dd6a874fd/volumes" Feb 25 14:15:24 crc kubenswrapper[4815]: I0225 14:15:24.926447 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7fvlc"] Feb 25 14:15:24 crc kubenswrapper[4815]: E0225 14:15:24.927474 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cc0d985-5b1d-4756-ac53-be6138bb46ba" containerName="collect-profiles" Feb 25 14:15:24 crc kubenswrapper[4815]: I0225 14:15:24.927491 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cc0d985-5b1d-4756-ac53-be6138bb46ba" containerName="collect-profiles" Feb 25 14:15:24 crc kubenswrapper[4815]: I0225 14:15:24.927946 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cc0d985-5b1d-4756-ac53-be6138bb46ba" containerName="collect-profiles" Feb 25 14:15:24 crc kubenswrapper[4815]: I0225 14:15:24.929593 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fvlc" Feb 25 14:15:24 crc kubenswrapper[4815]: I0225 14:15:24.951609 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fvlc"] Feb 25 14:15:25 crc kubenswrapper[4815]: I0225 14:15:25.108853 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32d69247-675b-4fae-9787-877e631ead8b-utilities\") pod \"redhat-marketplace-7fvlc\" (UID: \"32d69247-675b-4fae-9787-877e631ead8b\") " pod="openshift-marketplace/redhat-marketplace-7fvlc" Feb 25 14:15:25 crc kubenswrapper[4815]: I0225 14:15:25.108951 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32d69247-675b-4fae-9787-877e631ead8b-catalog-content\") pod \"redhat-marketplace-7fvlc\" (UID: \"32d69247-675b-4fae-9787-877e631ead8b\") " pod="openshift-marketplace/redhat-marketplace-7fvlc" Feb 25 14:15:25 crc kubenswrapper[4815]: I0225 14:15:25.109015 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntw87\" (UniqueName: \"kubernetes.io/projected/32d69247-675b-4fae-9787-877e631ead8b-kube-api-access-ntw87\") pod \"redhat-marketplace-7fvlc\" (UID: \"32d69247-675b-4fae-9787-877e631ead8b\") " pod="openshift-marketplace/redhat-marketplace-7fvlc" Feb 25 14:15:25 crc kubenswrapper[4815]: I0225 14:15:25.211006 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32d69247-675b-4fae-9787-877e631ead8b-utilities\") pod \"redhat-marketplace-7fvlc\" (UID: \"32d69247-675b-4fae-9787-877e631ead8b\") " pod="openshift-marketplace/redhat-marketplace-7fvlc" Feb 25 14:15:25 crc kubenswrapper[4815]: I0225 14:15:25.211075 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32d69247-675b-4fae-9787-877e631ead8b-catalog-content\") pod \"redhat-marketplace-7fvlc\" (UID: \"32d69247-675b-4fae-9787-877e631ead8b\") " pod="openshift-marketplace/redhat-marketplace-7fvlc" Feb 25 14:15:25 crc kubenswrapper[4815]: I0225 14:15:25.211128 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntw87\" (UniqueName: \"kubernetes.io/projected/32d69247-675b-4fae-9787-877e631ead8b-kube-api-access-ntw87\") pod \"redhat-marketplace-7fvlc\" (UID: \"32d69247-675b-4fae-9787-877e631ead8b\") " pod="openshift-marketplace/redhat-marketplace-7fvlc" Feb 25 14:15:25 crc kubenswrapper[4815]: I0225 14:15:25.211458 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32d69247-675b-4fae-9787-877e631ead8b-utilities\") pod \"redhat-marketplace-7fvlc\" (UID: \"32d69247-675b-4fae-9787-877e631ead8b\") " pod="openshift-marketplace/redhat-marketplace-7fvlc" Feb 25 14:15:25 crc kubenswrapper[4815]: I0225 14:15:25.211575 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32d69247-675b-4fae-9787-877e631ead8b-catalog-content\") pod \"redhat-marketplace-7fvlc\" (UID: \"32d69247-675b-4fae-9787-877e631ead8b\") " pod="openshift-marketplace/redhat-marketplace-7fvlc" Feb 25 14:15:25 crc kubenswrapper[4815]: I0225 14:15:25.230440 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntw87\" (UniqueName: \"kubernetes.io/projected/32d69247-675b-4fae-9787-877e631ead8b-kube-api-access-ntw87\") pod \"redhat-marketplace-7fvlc\" (UID: \"32d69247-675b-4fae-9787-877e631ead8b\") " pod="openshift-marketplace/redhat-marketplace-7fvlc" Feb 25 14:15:25 crc kubenswrapper[4815]: I0225 14:15:25.268464 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fvlc" Feb 25 14:15:25 crc kubenswrapper[4815]: I0225 14:15:25.813037 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fvlc"] Feb 25 14:15:26 crc kubenswrapper[4815]: I0225 14:15:26.403341 4815 generic.go:334] "Generic (PLEG): container finished" podID="32d69247-675b-4fae-9787-877e631ead8b" containerID="6402a9b6fd175c1056d55f27437fe97e94ed5a7e51119f10f43c8d03f954360a" exitCode=0 Feb 25 14:15:26 crc kubenswrapper[4815]: I0225 14:15:26.403463 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fvlc" event={"ID":"32d69247-675b-4fae-9787-877e631ead8b","Type":"ContainerDied","Data":"6402a9b6fd175c1056d55f27437fe97e94ed5a7e51119f10f43c8d03f954360a"} Feb 25 14:15:26 crc kubenswrapper[4815]: I0225 14:15:26.403672 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fvlc" event={"ID":"32d69247-675b-4fae-9787-877e631ead8b","Type":"ContainerStarted","Data":"3d6ef211eba303f43fa2e567982f0d5d2285a793c74b4705ac32c860b6873ab1"} Feb 25 14:15:27 crc kubenswrapper[4815]: I0225 14:15:27.413558 4815 generic.go:334] "Generic (PLEG): container finished" podID="32d69247-675b-4fae-9787-877e631ead8b" containerID="d955cd8358b123fe02cbdb0862ad59af1091f1aeb7544655fb2397abc1a9ea7a" exitCode=0 Feb 25 14:15:27 crc kubenswrapper[4815]: I0225 14:15:27.413611 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fvlc" event={"ID":"32d69247-675b-4fae-9787-877e631ead8b","Type":"ContainerDied","Data":"d955cd8358b123fe02cbdb0862ad59af1091f1aeb7544655fb2397abc1a9ea7a"} Feb 25 14:15:28 crc kubenswrapper[4815]: I0225 14:15:28.426027 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fvlc" event={"ID":"32d69247-675b-4fae-9787-877e631ead8b","Type":"ContainerStarted","Data":"a27d931a7d69482109035eebac11f8cc078ec6a80db3e49fec63ae3e547c56fe"} Feb 25 14:15:28 crc kubenswrapper[4815]: I0225 14:15:28.457821 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7fvlc" podStartSLOduration=3.050641144 podStartE2EDuration="4.457801588s" podCreationTimestamp="2026-02-25 14:15:24 +0000 UTC" firstStartedPulling="2026-02-25 14:15:26.407057337 +0000 UTC m=+3284.208155391" lastFinishedPulling="2026-02-25 14:15:27.814217791 +0000 UTC m=+3285.615315835" observedRunningTime="2026-02-25 14:15:28.448866618 +0000 UTC m=+3286.249964682" watchObservedRunningTime="2026-02-25 14:15:28.457801588 +0000 UTC m=+3286.258899642" Feb 25 14:15:33 crc kubenswrapper[4815]: I0225 14:15:33.475278 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-twd5c/must-gather-v7wxh"] Feb 25 14:15:33 crc kubenswrapper[4815]: I0225 14:15:33.486328 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-twd5c/must-gather-v7wxh" Feb 25 14:15:33 crc kubenswrapper[4815]: I0225 14:15:33.489330 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-twd5c"/"default-dockercfg-m6fj4" Feb 25 14:15:33 crc kubenswrapper[4815]: I0225 14:15:33.489399 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-twd5c"/"kube-root-ca.crt" Feb 25 14:15:33 crc kubenswrapper[4815]: I0225 14:15:33.503776 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-twd5c"/"openshift-service-ca.crt" Feb 25 14:15:33 crc kubenswrapper[4815]: I0225 14:15:33.523794 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-twd5c/must-gather-v7wxh"] Feb 25 14:15:33 crc kubenswrapper[4815]: I0225 14:15:33.601781 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szbq8\" (UniqueName: \"kubernetes.io/projected/8da95553-a7bd-4cd4-84ea-10a8acf68ced-kube-api-access-szbq8\") pod \"must-gather-v7wxh\" (UID: \"8da95553-a7bd-4cd4-84ea-10a8acf68ced\") " pod="openshift-must-gather-twd5c/must-gather-v7wxh" Feb 25 14:15:33 crc kubenswrapper[4815]: I0225 14:15:33.601838 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8da95553-a7bd-4cd4-84ea-10a8acf68ced-must-gather-output\") pod \"must-gather-v7wxh\" (UID: \"8da95553-a7bd-4cd4-84ea-10a8acf68ced\") " pod="openshift-must-gather-twd5c/must-gather-v7wxh" Feb 25 14:15:33 crc kubenswrapper[4815]: I0225 14:15:33.704128 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szbq8\" (UniqueName: \"kubernetes.io/projected/8da95553-a7bd-4cd4-84ea-10a8acf68ced-kube-api-access-szbq8\") pod \"must-gather-v7wxh\" (UID: \"8da95553-a7bd-4cd4-84ea-10a8acf68ced\") " pod="openshift-must-gather-twd5c/must-gather-v7wxh" Feb 25 14:15:33 crc kubenswrapper[4815]: I0225 14:15:33.704180 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8da95553-a7bd-4cd4-84ea-10a8acf68ced-must-gather-output\") pod \"must-gather-v7wxh\" (UID: \"8da95553-a7bd-4cd4-84ea-10a8acf68ced\") " pod="openshift-must-gather-twd5c/must-gather-v7wxh" Feb 25 14:15:33 crc kubenswrapper[4815]: I0225 14:15:33.705100 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8da95553-a7bd-4cd4-84ea-10a8acf68ced-must-gather-output\") pod \"must-gather-v7wxh\" (UID: \"8da95553-a7bd-4cd4-84ea-10a8acf68ced\") " pod="openshift-must-gather-twd5c/must-gather-v7wxh" Feb 25 14:15:33 crc kubenswrapper[4815]: I0225 14:15:33.724275 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szbq8\" (UniqueName: \"kubernetes.io/projected/8da95553-a7bd-4cd4-84ea-10a8acf68ced-kube-api-access-szbq8\") pod \"must-gather-v7wxh\" (UID: \"8da95553-a7bd-4cd4-84ea-10a8acf68ced\") " pod="openshift-must-gather-twd5c/must-gather-v7wxh" Feb 25 14:15:33 crc kubenswrapper[4815]: I0225 14:15:33.810217 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-twd5c/must-gather-v7wxh" Feb 25 14:15:34 crc kubenswrapper[4815]: I0225 14:15:34.282853 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-twd5c/must-gather-v7wxh"] Feb 25 14:15:34 crc kubenswrapper[4815]: I0225 14:15:34.493052 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-twd5c/must-gather-v7wxh" event={"ID":"8da95553-a7bd-4cd4-84ea-10a8acf68ced","Type":"ContainerStarted","Data":"71f38b05d5613754b12d2c08ed1c575ccf773b9f5961bf53cd7fb0b841c3fbc7"} Feb 25 14:15:35 crc kubenswrapper[4815]: I0225 14:15:35.269205 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7fvlc" Feb 25 14:15:35 crc kubenswrapper[4815]: I0225 14:15:35.269678 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7fvlc" Feb 25 14:15:35 crc kubenswrapper[4815]: I0225 14:15:35.321763 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7fvlc" Feb 25 14:15:35 crc kubenswrapper[4815]: I0225 14:15:35.552492 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7fvlc" Feb 25 14:15:35 crc kubenswrapper[4815]: I0225 14:15:35.617610 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fvlc"] Feb 25 14:15:37 crc kubenswrapper[4815]: I0225 14:15:37.534356 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7fvlc" podUID="32d69247-675b-4fae-9787-877e631ead8b" containerName="registry-server" containerID="cri-o://a27d931a7d69482109035eebac11f8cc078ec6a80db3e49fec63ae3e547c56fe" gracePeriod=2 Feb 25 14:15:38 crc kubenswrapper[4815]: I0225 14:15:38.545086 4815 generic.go:334] "Generic (PLEG): container finished" podID="32d69247-675b-4fae-9787-877e631ead8b" containerID="a27d931a7d69482109035eebac11f8cc078ec6a80db3e49fec63ae3e547c56fe" exitCode=0 Feb 25 14:15:38 crc kubenswrapper[4815]: I0225 14:15:38.545164 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fvlc" event={"ID":"32d69247-675b-4fae-9787-877e631ead8b","Type":"ContainerDied","Data":"a27d931a7d69482109035eebac11f8cc078ec6a80db3e49fec63ae3e547c56fe"} Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.526023 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fvlc" Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.583068 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-twd5c/must-gather-v7wxh" event={"ID":"8da95553-a7bd-4cd4-84ea-10a8acf68ced","Type":"ContainerStarted","Data":"6e262e4a7a08ff1daa2e69e8c3427adfc036dfff10ae58f297932ffa13e89dd9"} Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.589087 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fvlc" event={"ID":"32d69247-675b-4fae-9787-877e631ead8b","Type":"ContainerDied","Data":"3d6ef211eba303f43fa2e567982f0d5d2285a793c74b4705ac32c860b6873ab1"} Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.589135 4815 scope.go:117] "RemoveContainer" containerID="a27d931a7d69482109035eebac11f8cc078ec6a80db3e49fec63ae3e547c56fe" Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.589299 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fvlc" Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.590416 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32d69247-675b-4fae-9787-877e631ead8b-utilities\") pod \"32d69247-675b-4fae-9787-877e631ead8b\" (UID: \"32d69247-675b-4fae-9787-877e631ead8b\") " Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.590466 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32d69247-675b-4fae-9787-877e631ead8b-catalog-content\") pod \"32d69247-675b-4fae-9787-877e631ead8b\" (UID: \"32d69247-675b-4fae-9787-877e631ead8b\") " Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.590549 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntw87\" (UniqueName: \"kubernetes.io/projected/32d69247-675b-4fae-9787-877e631ead8b-kube-api-access-ntw87\") pod \"32d69247-675b-4fae-9787-877e631ead8b\" (UID: \"32d69247-675b-4fae-9787-877e631ead8b\") " Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.591499 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32d69247-675b-4fae-9787-877e631ead8b-utilities" (OuterVolumeSpecName: "utilities") pod "32d69247-675b-4fae-9787-877e631ead8b" (UID: "32d69247-675b-4fae-9787-877e631ead8b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.598217 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32d69247-675b-4fae-9787-877e631ead8b-kube-api-access-ntw87" (OuterVolumeSpecName: "kube-api-access-ntw87") pod "32d69247-675b-4fae-9787-877e631ead8b" (UID: "32d69247-675b-4fae-9787-877e631ead8b"). InnerVolumeSpecName "kube-api-access-ntw87". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.623857 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32d69247-675b-4fae-9787-877e631ead8b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "32d69247-675b-4fae-9787-877e631ead8b" (UID: "32d69247-675b-4fae-9787-877e631ead8b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.625802 4815 scope.go:117] "RemoveContainer" containerID="d955cd8358b123fe02cbdb0862ad59af1091f1aeb7544655fb2397abc1a9ea7a" Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.650097 4815 scope.go:117] "RemoveContainer" containerID="6402a9b6fd175c1056d55f27437fe97e94ed5a7e51119f10f43c8d03f954360a" Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.693097 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32d69247-675b-4fae-9787-877e631ead8b-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.693130 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32d69247-675b-4fae-9787-877e631ead8b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.693142 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntw87\" (UniqueName: \"kubernetes.io/projected/32d69247-675b-4fae-9787-877e631ead8b-kube-api-access-ntw87\") on node \"crc\" DevicePath \"\"" Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.927724 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fvlc"] Feb 25 14:15:41 crc kubenswrapper[4815]: I0225 14:15:41.937919 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fvlc"] Feb 25 14:15:42 crc kubenswrapper[4815]: I0225 14:15:42.600794 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-twd5c/must-gather-v7wxh" event={"ID":"8da95553-a7bd-4cd4-84ea-10a8acf68ced","Type":"ContainerStarted","Data":"11f64d995172c2029009adb72beb11cd88f94fd2d23a9d47d1839d0c5bc25625"} Feb 25 14:15:42 crc kubenswrapper[4815]: I0225 14:15:42.627342 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-twd5c/must-gather-v7wxh" podStartSLOduration=2.770711728 podStartE2EDuration="9.627320613s" podCreationTimestamp="2026-02-25 14:15:33 +0000 UTC" firstStartedPulling="2026-02-25 14:15:34.281074378 +0000 UTC m=+3292.082172452" lastFinishedPulling="2026-02-25 14:15:41.137683273 +0000 UTC m=+3298.938781337" observedRunningTime="2026-02-25 14:15:42.621755529 +0000 UTC m=+3300.422853883" watchObservedRunningTime="2026-02-25 14:15:42.627320613 +0000 UTC m=+3300.428418667" Feb 25 14:15:42 crc kubenswrapper[4815]: I0225 14:15:42.949246 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32d69247-675b-4fae-9787-877e631ead8b" path="/var/lib/kubelet/pods/32d69247-675b-4fae-9787-877e631ead8b/volumes" Feb 25 14:15:45 crc kubenswrapper[4815]: I0225 14:15:45.100523 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-twd5c/crc-debug-9fz5t"] Feb 25 14:15:45 crc kubenswrapper[4815]: E0225 14:15:45.101455 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32d69247-675b-4fae-9787-877e631ead8b" containerName="extract-content" Feb 25 14:15:45 crc kubenswrapper[4815]: I0225 14:15:45.101471 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="32d69247-675b-4fae-9787-877e631ead8b" containerName="extract-content" Feb 25 14:15:45 crc kubenswrapper[4815]: E0225 14:15:45.101479 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32d69247-675b-4fae-9787-877e631ead8b" containerName="registry-server" Feb 25 14:15:45 crc kubenswrapper[4815]: I0225 14:15:45.101485 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="32d69247-675b-4fae-9787-877e631ead8b" containerName="registry-server" Feb 25 14:15:45 crc kubenswrapper[4815]: E0225 14:15:45.101497 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32d69247-675b-4fae-9787-877e631ead8b" containerName="extract-utilities" Feb 25 14:15:45 crc kubenswrapper[4815]: I0225 14:15:45.101525 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="32d69247-675b-4fae-9787-877e631ead8b" containerName="extract-utilities" Feb 25 14:15:45 crc kubenswrapper[4815]: I0225 14:15:45.101704 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="32d69247-675b-4fae-9787-877e631ead8b" containerName="registry-server" Feb 25 14:15:45 crc kubenswrapper[4815]: I0225 14:15:45.102322 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-twd5c/crc-debug-9fz5t" Feb 25 14:15:45 crc kubenswrapper[4815]: I0225 14:15:45.160246 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8b4a70d0-ee7f-44b3-94d9-629e4738ce50-host\") pod \"crc-debug-9fz5t\" (UID: \"8b4a70d0-ee7f-44b3-94d9-629e4738ce50\") " pod="openshift-must-gather-twd5c/crc-debug-9fz5t" Feb 25 14:15:45 crc kubenswrapper[4815]: I0225 14:15:45.160309 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pzp6\" (UniqueName: \"kubernetes.io/projected/8b4a70d0-ee7f-44b3-94d9-629e4738ce50-kube-api-access-2pzp6\") pod \"crc-debug-9fz5t\" (UID: \"8b4a70d0-ee7f-44b3-94d9-629e4738ce50\") " pod="openshift-must-gather-twd5c/crc-debug-9fz5t" Feb 25 14:15:45 crc kubenswrapper[4815]: I0225 14:15:45.262977 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8b4a70d0-ee7f-44b3-94d9-629e4738ce50-host\") pod \"crc-debug-9fz5t\" (UID: \"8b4a70d0-ee7f-44b3-94d9-629e4738ce50\") " pod="openshift-must-gather-twd5c/crc-debug-9fz5t" Feb 25 14:15:45 crc kubenswrapper[4815]: I0225 14:15:45.263367 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pzp6\" (UniqueName: \"kubernetes.io/projected/8b4a70d0-ee7f-44b3-94d9-629e4738ce50-kube-api-access-2pzp6\") pod \"crc-debug-9fz5t\" (UID: \"8b4a70d0-ee7f-44b3-94d9-629e4738ce50\") " pod="openshift-must-gather-twd5c/crc-debug-9fz5t" Feb 25 14:15:45 crc kubenswrapper[4815]: I0225 14:15:45.263149 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8b4a70d0-ee7f-44b3-94d9-629e4738ce50-host\") pod \"crc-debug-9fz5t\" (UID: \"8b4a70d0-ee7f-44b3-94d9-629e4738ce50\") " pod="openshift-must-gather-twd5c/crc-debug-9fz5t" Feb 25 14:15:45 crc kubenswrapper[4815]: I0225 14:15:45.298241 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pzp6\" (UniqueName: \"kubernetes.io/projected/8b4a70d0-ee7f-44b3-94d9-629e4738ce50-kube-api-access-2pzp6\") pod \"crc-debug-9fz5t\" (UID: \"8b4a70d0-ee7f-44b3-94d9-629e4738ce50\") " pod="openshift-must-gather-twd5c/crc-debug-9fz5t" Feb 25 14:15:45 crc kubenswrapper[4815]: I0225 14:15:45.419429 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-twd5c/crc-debug-9fz5t" Feb 25 14:15:45 crc kubenswrapper[4815]: W0225 14:15:45.467603 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b4a70d0_ee7f_44b3_94d9_629e4738ce50.slice/crio-63132ebd3a7d39c1db43624593a2ae723e3430a4d6a8411774f80d74c3161345 WatchSource:0}: Error finding container 63132ebd3a7d39c1db43624593a2ae723e3430a4d6a8411774f80d74c3161345: Status 404 returned error can't find the container with id 63132ebd3a7d39c1db43624593a2ae723e3430a4d6a8411774f80d74c3161345 Feb 25 14:15:45 crc kubenswrapper[4815]: I0225 14:15:45.630339 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-twd5c/crc-debug-9fz5t" event={"ID":"8b4a70d0-ee7f-44b3-94d9-629e4738ce50","Type":"ContainerStarted","Data":"63132ebd3a7d39c1db43624593a2ae723e3430a4d6a8411774f80d74c3161345"} Feb 25 14:15:49 crc kubenswrapper[4815]: I0225 14:15:49.713195 4815 scope.go:117] "RemoveContainer" containerID="cc3f45f46c471f80b48f0128822f6de4c029341b91547818a21a8764a1ad026a" Feb 25 14:15:57 crc kubenswrapper[4815]: I0225 14:15:57.736732 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-twd5c/crc-debug-9fz5t" event={"ID":"8b4a70d0-ee7f-44b3-94d9-629e4738ce50","Type":"ContainerStarted","Data":"2b8f0cbbadd1b0bb08e8907e0d677eef59504fdca28280b5bf3a5f296918cfb1"} Feb 25 14:15:57 crc kubenswrapper[4815]: I0225 14:15:57.753804 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-twd5c/crc-debug-9fz5t" podStartSLOduration=1.212360856 podStartE2EDuration="12.753787479s" podCreationTimestamp="2026-02-25 14:15:45 +0000 UTC" firstStartedPulling="2026-02-25 14:15:45.470039193 +0000 UTC m=+3303.271137247" lastFinishedPulling="2026-02-25 14:15:57.011465826 +0000 UTC m=+3314.812563870" observedRunningTime="2026-02-25 14:15:57.748924537 +0000 UTC m=+3315.550022591" watchObservedRunningTime="2026-02-25 14:15:57.753787479 +0000 UTC m=+3315.554885543" Feb 25 14:16:00 crc kubenswrapper[4815]: I0225 14:16:00.139322 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533816-pj8lb"] Feb 25 14:16:00 crc kubenswrapper[4815]: I0225 14:16:00.141467 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533816-pj8lb" Feb 25 14:16:00 crc kubenswrapper[4815]: I0225 14:16:00.148187 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 14:16:00 crc kubenswrapper[4815]: I0225 14:16:00.148191 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 14:16:00 crc kubenswrapper[4815]: I0225 14:16:00.148788 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 14:16:00 crc kubenswrapper[4815]: I0225 14:16:00.153871 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533816-pj8lb"] Feb 25 14:16:00 crc kubenswrapper[4815]: I0225 14:16:00.158644 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzr72\" (UniqueName: \"kubernetes.io/projected/5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1-kube-api-access-lzr72\") pod \"auto-csr-approver-29533816-pj8lb\" (UID: \"5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1\") " pod="openshift-infra/auto-csr-approver-29533816-pj8lb" Feb 25 14:16:00 crc kubenswrapper[4815]: I0225 14:16:00.260916 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzr72\" (UniqueName: \"kubernetes.io/projected/5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1-kube-api-access-lzr72\") pod \"auto-csr-approver-29533816-pj8lb\" (UID: \"5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1\") " pod="openshift-infra/auto-csr-approver-29533816-pj8lb" Feb 25 14:16:00 crc kubenswrapper[4815]: I0225 14:16:00.285881 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzr72\" (UniqueName: \"kubernetes.io/projected/5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1-kube-api-access-lzr72\") pod \"auto-csr-approver-29533816-pj8lb\" (UID: \"5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1\") " pod="openshift-infra/auto-csr-approver-29533816-pj8lb" Feb 25 14:16:00 crc kubenswrapper[4815]: I0225 14:16:00.460726 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533816-pj8lb" Feb 25 14:16:00 crc kubenswrapper[4815]: I0225 14:16:00.923998 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533816-pj8lb"] Feb 25 14:16:01 crc kubenswrapper[4815]: I0225 14:16:01.784640 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533816-pj8lb" event={"ID":"5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1","Type":"ContainerStarted","Data":"3809417e494d0c810f5099b6c1e664d64997512b90e632b7a606edb95fb716c5"} Feb 25 14:16:04 crc kubenswrapper[4815]: I0225 14:16:04.822384 4815 generic.go:334] "Generic (PLEG): container finished" podID="5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1" containerID="1c9a31b8117a83d4c038cab9f9f6330dba60646d171fc2e6dba06da20c2a0966" exitCode=0 Feb 25 14:16:04 crc kubenswrapper[4815]: I0225 14:16:04.822478 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533816-pj8lb" event={"ID":"5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1","Type":"ContainerDied","Data":"1c9a31b8117a83d4c038cab9f9f6330dba60646d171fc2e6dba06da20c2a0966"} Feb 25 14:16:06 crc kubenswrapper[4815]: I0225 14:16:06.189965 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533816-pj8lb" Feb 25 14:16:06 crc kubenswrapper[4815]: I0225 14:16:06.374765 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzr72\" (UniqueName: \"kubernetes.io/projected/5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1-kube-api-access-lzr72\") pod \"5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1\" (UID: \"5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1\") " Feb 25 14:16:06 crc kubenswrapper[4815]: I0225 14:16:06.381735 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1-kube-api-access-lzr72" (OuterVolumeSpecName: "kube-api-access-lzr72") pod "5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1" (UID: "5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1"). InnerVolumeSpecName "kube-api-access-lzr72". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:16:06 crc kubenswrapper[4815]: I0225 14:16:06.477883 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzr72\" (UniqueName: \"kubernetes.io/projected/5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1-kube-api-access-lzr72\") on node \"crc\" DevicePath \"\"" Feb 25 14:16:06 crc kubenswrapper[4815]: I0225 14:16:06.842421 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533816-pj8lb" event={"ID":"5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1","Type":"ContainerDied","Data":"3809417e494d0c810f5099b6c1e664d64997512b90e632b7a606edb95fb716c5"} Feb 25 14:16:06 crc kubenswrapper[4815]: I0225 14:16:06.842771 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3809417e494d0c810f5099b6c1e664d64997512b90e632b7a606edb95fb716c5" Feb 25 14:16:06 crc kubenswrapper[4815]: I0225 14:16:06.842644 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533816-pj8lb" Feb 25 14:16:07 crc kubenswrapper[4815]: I0225 14:16:07.278039 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533810-j9ssf"] Feb 25 14:16:07 crc kubenswrapper[4815]: I0225 14:16:07.288651 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533810-j9ssf"] Feb 25 14:16:08 crc kubenswrapper[4815]: I0225 14:16:08.953680 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f5569c7-1d59-4aec-9f95-a27707e08b52" path="/var/lib/kubelet/pods/4f5569c7-1d59-4aec-9f95-a27707e08b52/volumes" Feb 25 14:16:36 crc kubenswrapper[4815]: I0225 14:16:36.113297 4815 generic.go:334] "Generic (PLEG): container finished" podID="8b4a70d0-ee7f-44b3-94d9-629e4738ce50" containerID="2b8f0cbbadd1b0bb08e8907e0d677eef59504fdca28280b5bf3a5f296918cfb1" exitCode=0 Feb 25 14:16:36 crc kubenswrapper[4815]: I0225 14:16:36.113477 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-twd5c/crc-debug-9fz5t" event={"ID":"8b4a70d0-ee7f-44b3-94d9-629e4738ce50","Type":"ContainerDied","Data":"2b8f0cbbadd1b0bb08e8907e0d677eef59504fdca28280b5bf3a5f296918cfb1"} Feb 25 14:16:37 crc kubenswrapper[4815]: I0225 14:16:37.229029 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-twd5c/crc-debug-9fz5t" Feb 25 14:16:37 crc kubenswrapper[4815]: I0225 14:16:37.263682 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-twd5c/crc-debug-9fz5t"] Feb 25 14:16:37 crc kubenswrapper[4815]: I0225 14:16:37.268656 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pzp6\" (UniqueName: \"kubernetes.io/projected/8b4a70d0-ee7f-44b3-94d9-629e4738ce50-kube-api-access-2pzp6\") pod \"8b4a70d0-ee7f-44b3-94d9-629e4738ce50\" (UID: \"8b4a70d0-ee7f-44b3-94d9-629e4738ce50\") " Feb 25 14:16:37 crc kubenswrapper[4815]: I0225 14:16:37.268753 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8b4a70d0-ee7f-44b3-94d9-629e4738ce50-host\") pod \"8b4a70d0-ee7f-44b3-94d9-629e4738ce50\" (UID: \"8b4a70d0-ee7f-44b3-94d9-629e4738ce50\") " Feb 25 14:16:37 crc kubenswrapper[4815]: I0225 14:16:37.268838 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b4a70d0-ee7f-44b3-94d9-629e4738ce50-host" (OuterVolumeSpecName: "host") pod "8b4a70d0-ee7f-44b3-94d9-629e4738ce50" (UID: "8b4a70d0-ee7f-44b3-94d9-629e4738ce50"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 14:16:37 crc kubenswrapper[4815]: I0225 14:16:37.269077 4815 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8b4a70d0-ee7f-44b3-94d9-629e4738ce50-host\") on node \"crc\" DevicePath \"\"" Feb 25 14:16:37 crc kubenswrapper[4815]: I0225 14:16:37.273633 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-twd5c/crc-debug-9fz5t"] Feb 25 14:16:37 crc kubenswrapper[4815]: I0225 14:16:37.274571 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b4a70d0-ee7f-44b3-94d9-629e4738ce50-kube-api-access-2pzp6" (OuterVolumeSpecName: "kube-api-access-2pzp6") pod "8b4a70d0-ee7f-44b3-94d9-629e4738ce50" (UID: "8b4a70d0-ee7f-44b3-94d9-629e4738ce50"). InnerVolumeSpecName "kube-api-access-2pzp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:16:37 crc kubenswrapper[4815]: I0225 14:16:37.371264 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pzp6\" (UniqueName: \"kubernetes.io/projected/8b4a70d0-ee7f-44b3-94d9-629e4738ce50-kube-api-access-2pzp6\") on node \"crc\" DevicePath \"\"" Feb 25 14:16:38 crc kubenswrapper[4815]: I0225 14:16:38.135973 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63132ebd3a7d39c1db43624593a2ae723e3430a4d6a8411774f80d74c3161345" Feb 25 14:16:38 crc kubenswrapper[4815]: I0225 14:16:38.136019 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-twd5c/crc-debug-9fz5t" Feb 25 14:16:38 crc kubenswrapper[4815]: I0225 14:16:38.460720 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-twd5c/crc-debug-99jwh"] Feb 25 14:16:38 crc kubenswrapper[4815]: E0225 14:16:38.461183 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1" containerName="oc" Feb 25 14:16:38 crc kubenswrapper[4815]: I0225 14:16:38.461199 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1" containerName="oc" Feb 25 14:16:38 crc kubenswrapper[4815]: E0225 14:16:38.461269 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b4a70d0-ee7f-44b3-94d9-629e4738ce50" containerName="container-00" Feb 25 14:16:38 crc kubenswrapper[4815]: I0225 14:16:38.461282 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b4a70d0-ee7f-44b3-94d9-629e4738ce50" containerName="container-00" Feb 25 14:16:38 crc kubenswrapper[4815]: I0225 14:16:38.461917 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1" containerName="oc" Feb 25 14:16:38 crc kubenswrapper[4815]: I0225 14:16:38.461967 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b4a70d0-ee7f-44b3-94d9-629e4738ce50" containerName="container-00" Feb 25 14:16:38 crc kubenswrapper[4815]: I0225 14:16:38.462994 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-twd5c/crc-debug-99jwh" Feb 25 14:16:38 crc kubenswrapper[4815]: I0225 14:16:38.491676 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/303851ee-6634-4144-84af-69eae85586ed-host\") pod \"crc-debug-99jwh\" (UID: \"303851ee-6634-4144-84af-69eae85586ed\") " pod="openshift-must-gather-twd5c/crc-debug-99jwh" Feb 25 14:16:38 crc kubenswrapper[4815]: I0225 14:16:38.492009 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv7dl\" (UniqueName: \"kubernetes.io/projected/303851ee-6634-4144-84af-69eae85586ed-kube-api-access-bv7dl\") pod \"crc-debug-99jwh\" (UID: \"303851ee-6634-4144-84af-69eae85586ed\") " pod="openshift-must-gather-twd5c/crc-debug-99jwh" Feb 25 14:16:38 crc kubenswrapper[4815]: I0225 14:16:38.593549 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv7dl\" (UniqueName: \"kubernetes.io/projected/303851ee-6634-4144-84af-69eae85586ed-kube-api-access-bv7dl\") pod \"crc-debug-99jwh\" (UID: \"303851ee-6634-4144-84af-69eae85586ed\") " pod="openshift-must-gather-twd5c/crc-debug-99jwh" Feb 25 14:16:38 crc kubenswrapper[4815]: I0225 14:16:38.593865 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/303851ee-6634-4144-84af-69eae85586ed-host\") pod \"crc-debug-99jwh\" (UID: \"303851ee-6634-4144-84af-69eae85586ed\") " pod="openshift-must-gather-twd5c/crc-debug-99jwh" Feb 25 14:16:38 crc kubenswrapper[4815]: I0225 14:16:38.594006 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/303851ee-6634-4144-84af-69eae85586ed-host\") pod \"crc-debug-99jwh\" (UID: \"303851ee-6634-4144-84af-69eae85586ed\") " pod="openshift-must-gather-twd5c/crc-debug-99jwh" Feb 25 14:16:38 crc kubenswrapper[4815]: I0225 14:16:38.613685 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv7dl\" (UniqueName: \"kubernetes.io/projected/303851ee-6634-4144-84af-69eae85586ed-kube-api-access-bv7dl\") pod \"crc-debug-99jwh\" (UID: \"303851ee-6634-4144-84af-69eae85586ed\") " pod="openshift-must-gather-twd5c/crc-debug-99jwh" Feb 25 14:16:38 crc kubenswrapper[4815]: I0225 14:16:38.785572 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-twd5c/crc-debug-99jwh" Feb 25 14:16:38 crc kubenswrapper[4815]: I0225 14:16:38.948437 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b4a70d0-ee7f-44b3-94d9-629e4738ce50" path="/var/lib/kubelet/pods/8b4a70d0-ee7f-44b3-94d9-629e4738ce50/volumes" Feb 25 14:16:39 crc kubenswrapper[4815]: I0225 14:16:39.146716 4815 generic.go:334] "Generic (PLEG): container finished" podID="303851ee-6634-4144-84af-69eae85586ed" containerID="bc38ccd47f6ad13c184bba624727bad0432abdafa683d560285c603596000696" exitCode=0 Feb 25 14:16:39 crc kubenswrapper[4815]: I0225 14:16:39.146760 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-twd5c/crc-debug-99jwh" event={"ID":"303851ee-6634-4144-84af-69eae85586ed","Type":"ContainerDied","Data":"bc38ccd47f6ad13c184bba624727bad0432abdafa683d560285c603596000696"} Feb 25 14:16:39 crc kubenswrapper[4815]: I0225 14:16:39.147024 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-twd5c/crc-debug-99jwh" event={"ID":"303851ee-6634-4144-84af-69eae85586ed","Type":"ContainerStarted","Data":"4dba07869baa759a0153df19024e117af7a9ea81c791b169ada5f13d22942329"} Feb 25 14:16:39 crc kubenswrapper[4815]: I0225 14:16:39.582498 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-twd5c/crc-debug-99jwh"] Feb 25 14:16:39 crc kubenswrapper[4815]: I0225 14:16:39.592864 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-twd5c/crc-debug-99jwh"] Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.252458 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-twd5c/crc-debug-99jwh" Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.320393 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv7dl\" (UniqueName: \"kubernetes.io/projected/303851ee-6634-4144-84af-69eae85586ed-kube-api-access-bv7dl\") pod \"303851ee-6634-4144-84af-69eae85586ed\" (UID: \"303851ee-6634-4144-84af-69eae85586ed\") " Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.320451 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/303851ee-6634-4144-84af-69eae85586ed-host\") pod \"303851ee-6634-4144-84af-69eae85586ed\" (UID: \"303851ee-6634-4144-84af-69eae85586ed\") " Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.320655 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/303851ee-6634-4144-84af-69eae85586ed-host" (OuterVolumeSpecName: "host") pod "303851ee-6634-4144-84af-69eae85586ed" (UID: "303851ee-6634-4144-84af-69eae85586ed"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.321030 4815 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/303851ee-6634-4144-84af-69eae85586ed-host\") on node \"crc\" DevicePath \"\"" Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.325907 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/303851ee-6634-4144-84af-69eae85586ed-kube-api-access-bv7dl" (OuterVolumeSpecName: "kube-api-access-bv7dl") pod "303851ee-6634-4144-84af-69eae85586ed" (UID: "303851ee-6634-4144-84af-69eae85586ed"). InnerVolumeSpecName "kube-api-access-bv7dl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.422780 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv7dl\" (UniqueName: \"kubernetes.io/projected/303851ee-6634-4144-84af-69eae85586ed-kube-api-access-bv7dl\") on node \"crc\" DevicePath \"\"" Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.771216 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-twd5c/crc-debug-5zjnj"] Feb 25 14:16:40 crc kubenswrapper[4815]: E0225 14:16:40.772399 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="303851ee-6634-4144-84af-69eae85586ed" containerName="container-00" Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.772484 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="303851ee-6634-4144-84af-69eae85586ed" containerName="container-00" Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.772733 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="303851ee-6634-4144-84af-69eae85586ed" containerName="container-00" Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.773680 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-twd5c/crc-debug-5zjnj" Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.830900 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9jj2\" (UniqueName: \"kubernetes.io/projected/c53cc96d-fb25-4a3e-874d-731604187c5a-kube-api-access-q9jj2\") pod \"crc-debug-5zjnj\" (UID: \"c53cc96d-fb25-4a3e-874d-731604187c5a\") " pod="openshift-must-gather-twd5c/crc-debug-5zjnj" Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.831114 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c53cc96d-fb25-4a3e-874d-731604187c5a-host\") pod \"crc-debug-5zjnj\" (UID: \"c53cc96d-fb25-4a3e-874d-731604187c5a\") " pod="openshift-must-gather-twd5c/crc-debug-5zjnj" Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.933352 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9jj2\" (UniqueName: \"kubernetes.io/projected/c53cc96d-fb25-4a3e-874d-731604187c5a-kube-api-access-q9jj2\") pod \"crc-debug-5zjnj\" (UID: \"c53cc96d-fb25-4a3e-874d-731604187c5a\") " pod="openshift-must-gather-twd5c/crc-debug-5zjnj" Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.933523 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c53cc96d-fb25-4a3e-874d-731604187c5a-host\") pod \"crc-debug-5zjnj\" (UID: \"c53cc96d-fb25-4a3e-874d-731604187c5a\") " pod="openshift-must-gather-twd5c/crc-debug-5zjnj" Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.933666 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c53cc96d-fb25-4a3e-874d-731604187c5a-host\") pod \"crc-debug-5zjnj\" (UID: \"c53cc96d-fb25-4a3e-874d-731604187c5a\") " pod="openshift-must-gather-twd5c/crc-debug-5zjnj" Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.951072 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="303851ee-6634-4144-84af-69eae85586ed" path="/var/lib/kubelet/pods/303851ee-6634-4144-84af-69eae85586ed/volumes" Feb 25 14:16:40 crc kubenswrapper[4815]: I0225 14:16:40.958597 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9jj2\" (UniqueName: \"kubernetes.io/projected/c53cc96d-fb25-4a3e-874d-731604187c5a-kube-api-access-q9jj2\") pod \"crc-debug-5zjnj\" (UID: \"c53cc96d-fb25-4a3e-874d-731604187c5a\") " pod="openshift-must-gather-twd5c/crc-debug-5zjnj" Feb 25 14:16:41 crc kubenswrapper[4815]: I0225 14:16:41.091190 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-twd5c/crc-debug-5zjnj" Feb 25 14:16:41 crc kubenswrapper[4815]: W0225 14:16:41.140440 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc53cc96d_fb25_4a3e_874d_731604187c5a.slice/crio-46e54dba398e3d57463b403e9b4d5e24bcbfcc2880bc19041f60e75e5c72c25c WatchSource:0}: Error finding container 46e54dba398e3d57463b403e9b4d5e24bcbfcc2880bc19041f60e75e5c72c25c: Status 404 returned error can't find the container with id 46e54dba398e3d57463b403e9b4d5e24bcbfcc2880bc19041f60e75e5c72c25c Feb 25 14:16:41 crc kubenswrapper[4815]: I0225 14:16:41.165447 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-twd5c/crc-debug-5zjnj" event={"ID":"c53cc96d-fb25-4a3e-874d-731604187c5a","Type":"ContainerStarted","Data":"46e54dba398e3d57463b403e9b4d5e24bcbfcc2880bc19041f60e75e5c72c25c"} Feb 25 14:16:41 crc kubenswrapper[4815]: I0225 14:16:41.167891 4815 scope.go:117] "RemoveContainer" containerID="bc38ccd47f6ad13c184bba624727bad0432abdafa683d560285c603596000696" Feb 25 14:16:41 crc kubenswrapper[4815]: I0225 14:16:41.167972 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-twd5c/crc-debug-99jwh" Feb 25 14:16:42 crc kubenswrapper[4815]: I0225 14:16:42.179836 4815 generic.go:334] "Generic (PLEG): container finished" podID="c53cc96d-fb25-4a3e-874d-731604187c5a" containerID="f504f42e10c1784ffb10a59453252885a980c55f175d6c4e60137f161c8431aa" exitCode=0 Feb 25 14:16:42 crc kubenswrapper[4815]: I0225 14:16:42.179909 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-twd5c/crc-debug-5zjnj" event={"ID":"c53cc96d-fb25-4a3e-874d-731604187c5a","Type":"ContainerDied","Data":"f504f42e10c1784ffb10a59453252885a980c55f175d6c4e60137f161c8431aa"} Feb 25 14:16:42 crc kubenswrapper[4815]: I0225 14:16:42.227718 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-twd5c/crc-debug-5zjnj"] Feb 25 14:16:42 crc kubenswrapper[4815]: I0225 14:16:42.247491 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-twd5c/crc-debug-5zjnj"] Feb 25 14:16:43 crc kubenswrapper[4815]: I0225 14:16:43.310985 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-twd5c/crc-debug-5zjnj" Feb 25 14:16:43 crc kubenswrapper[4815]: I0225 14:16:43.483495 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9jj2\" (UniqueName: \"kubernetes.io/projected/c53cc96d-fb25-4a3e-874d-731604187c5a-kube-api-access-q9jj2\") pod \"c53cc96d-fb25-4a3e-874d-731604187c5a\" (UID: \"c53cc96d-fb25-4a3e-874d-731604187c5a\") " Feb 25 14:16:43 crc kubenswrapper[4815]: I0225 14:16:43.483805 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c53cc96d-fb25-4a3e-874d-731604187c5a-host\") pod \"c53cc96d-fb25-4a3e-874d-731604187c5a\" (UID: \"c53cc96d-fb25-4a3e-874d-731604187c5a\") " Feb 25 14:16:43 crc kubenswrapper[4815]: I0225 14:16:43.483936 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c53cc96d-fb25-4a3e-874d-731604187c5a-host" (OuterVolumeSpecName: "host") pod "c53cc96d-fb25-4a3e-874d-731604187c5a" (UID: "c53cc96d-fb25-4a3e-874d-731604187c5a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 14:16:43 crc kubenswrapper[4815]: I0225 14:16:43.484384 4815 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c53cc96d-fb25-4a3e-874d-731604187c5a-host\") on node \"crc\" DevicePath \"\"" Feb 25 14:16:43 crc kubenswrapper[4815]: I0225 14:16:43.493820 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c53cc96d-fb25-4a3e-874d-731604187c5a-kube-api-access-q9jj2" (OuterVolumeSpecName: "kube-api-access-q9jj2") pod "c53cc96d-fb25-4a3e-874d-731604187c5a" (UID: "c53cc96d-fb25-4a3e-874d-731604187c5a"). InnerVolumeSpecName "kube-api-access-q9jj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:16:43 crc kubenswrapper[4815]: I0225 14:16:43.586315 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9jj2\" (UniqueName: \"kubernetes.io/projected/c53cc96d-fb25-4a3e-874d-731604187c5a-kube-api-access-q9jj2\") on node \"crc\" DevicePath \"\"" Feb 25 14:16:44 crc kubenswrapper[4815]: I0225 14:16:44.209211 4815 scope.go:117] "RemoveContainer" containerID="f504f42e10c1784ffb10a59453252885a980c55f175d6c4e60137f161c8431aa" Feb 25 14:16:44 crc kubenswrapper[4815]: I0225 14:16:44.209265 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-twd5c/crc-debug-5zjnj" Feb 25 14:16:44 crc kubenswrapper[4815]: I0225 14:16:44.950023 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c53cc96d-fb25-4a3e-874d-731604187c5a" path="/var/lib/kubelet/pods/c53cc96d-fb25-4a3e-874d-731604187c5a/volumes" Feb 25 14:16:49 crc kubenswrapper[4815]: I0225 14:16:49.798191 4815 scope.go:117] "RemoveContainer" containerID="ba245b6ab9f0ba822c6ae549e58c31c8928e482602ebdbb2db4363f34083044a" Feb 25 14:16:51 crc kubenswrapper[4815]: I0225 14:16:51.342048 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 14:16:51 crc kubenswrapper[4815]: I0225 14:16:51.342449 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 14:16:57 crc kubenswrapper[4815]: I0225 14:16:57.368836 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7fc6fdc978-tzrxw_cd01efd5-a5c6-449d-b539-f31454945163/barbican-api/0.log" Feb 25 14:16:57 crc kubenswrapper[4815]: I0225 14:16:57.489227 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7fc6fdc978-tzrxw_cd01efd5-a5c6-449d-b539-f31454945163/barbican-api-log/0.log" Feb 25 14:16:57 crc kubenswrapper[4815]: I0225 14:16:57.600578 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-b4dcd45c8-7j8v9_877d6ac2-0a5e-43bd-b387-0a2ba928b54f/barbican-keystone-listener-log/0.log" Feb 25 14:16:57 crc kubenswrapper[4815]: I0225 14:16:57.620200 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-b4dcd45c8-7j8v9_877d6ac2-0a5e-43bd-b387-0a2ba928b54f/barbican-keystone-listener/0.log" Feb 25 14:16:57 crc kubenswrapper[4815]: I0225 14:16:57.755302 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-77f7d79d79-6znrw_df0b5667-833f-40d6-8ef4-5c988adae0ba/barbican-worker/0.log" Feb 25 14:16:57 crc kubenswrapper[4815]: I0225 14:16:57.838259 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-77f7d79d79-6znrw_df0b5667-833f-40d6-8ef4-5c988adae0ba/barbican-worker-log/0.log" Feb 25 14:16:57 crc kubenswrapper[4815]: I0225 14:16:57.985265 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-nmjqz_3a6985a9-1674-41ff-b441-23a83eff2148/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 25 14:16:58 crc kubenswrapper[4815]: I0225 14:16:58.103186 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_98c6009e-1597-4edc-a7d9-3987e946bd57/ceilometer-central-agent/0.log" Feb 25 14:16:58 crc kubenswrapper[4815]: I0225 14:16:58.151168 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_98c6009e-1597-4edc-a7d9-3987e946bd57/ceilometer-notification-agent/0.log" Feb 25 14:16:58 crc kubenswrapper[4815]: I0225 14:16:58.199683 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_98c6009e-1597-4edc-a7d9-3987e946bd57/proxy-httpd/0.log" Feb 25 14:16:58 crc kubenswrapper[4815]: I0225 14:16:58.225144 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_98c6009e-1597-4edc-a7d9-3987e946bd57/sg-core/0.log" Feb 25 14:16:58 crc kubenswrapper[4815]: I0225 14:16:58.393638 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_fb136f80-7ca0-4315-a9f9-b36e161e6b0d/cinder-api/0.log" Feb 25 14:16:58 crc kubenswrapper[4815]: I0225 14:16:58.398423 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_fb136f80-7ca0-4315-a9f9-b36e161e6b0d/cinder-api-log/0.log" Feb 25 14:16:58 crc kubenswrapper[4815]: I0225 14:16:58.595149 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_64bb2d31-f562-4398-a0d6-c6bee9b9ba35/probe/0.log" Feb 25 14:16:58 crc kubenswrapper[4815]: I0225 14:16:58.600995 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_64bb2d31-f562-4398-a0d6-c6bee9b9ba35/cinder-scheduler/0.log" Feb 25 14:16:58 crc kubenswrapper[4815]: I0225 14:16:58.676543 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-dvgjt_52d4b958-a8c5-42ee-bd46-d541d47629cb/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 25 14:16:58 crc kubenswrapper[4815]: I0225 14:16:58.820081 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-fpslb_9f769896-45d8-45d0-8629-cee07d2cf4ea/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 25 14:16:59 crc kubenswrapper[4815]: I0225 14:16:59.027951 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-tft6z_ece2c0c0-fcc9-4919-aa49-4b3df08531e4/init/0.log" Feb 25 14:16:59 crc kubenswrapper[4815]: I0225 14:16:59.191640 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-tft6z_ece2c0c0-fcc9-4919-aa49-4b3df08531e4/init/0.log" Feb 25 14:16:59 crc kubenswrapper[4815]: I0225 14:16:59.238534 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-tft6z_ece2c0c0-fcc9-4919-aa49-4b3df08531e4/dnsmasq-dns/0.log" Feb 25 14:16:59 crc kubenswrapper[4815]: I0225 14:16:59.260894 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-4nj2v_9d6402d2-d86e-4529-a4bb-b6b26f9ed63c/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 25 14:16:59 crc kubenswrapper[4815]: I0225 14:16:59.425026 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f30dde44-ecd6-4908-b18e-1f9ba7f43fd7/glance-httpd/0.log" Feb 25 14:16:59 crc kubenswrapper[4815]: I0225 14:16:59.537713 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f30dde44-ecd6-4908-b18e-1f9ba7f43fd7/glance-log/0.log" Feb 25 14:16:59 crc kubenswrapper[4815]: I0225 14:16:59.615574 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_17cf15ab-9607-484e-b043-fffcd641dc99/glance-httpd/0.log" Feb 25 14:16:59 crc kubenswrapper[4815]: I0225 14:16:59.645385 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_17cf15ab-9607-484e-b043-fffcd641dc99/glance-log/0.log" Feb 25 14:16:59 crc kubenswrapper[4815]: I0225 14:16:59.917627 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7657655688-k9k6t_f10b8779-9d5e-427d-847c-3250816e89b9/horizon/0.log" Feb 25 14:17:00 crc kubenswrapper[4815]: I0225 14:17:00.146400 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-ttw7v_b70de6b1-ba0c-482d-a1d6-3c0604a374c1/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 25 14:17:00 crc kubenswrapper[4815]: I0225 14:17:00.180923 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7657655688-k9k6t_f10b8779-9d5e-427d-847c-3250816e89b9/horizon-log/0.log" Feb 25 14:17:00 crc kubenswrapper[4815]: I0225 14:17:00.212537 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-sgwbs_d6add50b-b148-4e9b-af2a-36571841b3c0/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 25 14:17:00 crc kubenswrapper[4815]: I0225 14:17:00.446718 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29533801-mmp7n_99ed322b-3f0d-4adc-8244-07af594b14e7/keystone-cron/0.log" Feb 25 14:17:00 crc kubenswrapper[4815]: I0225 14:17:00.535573 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6cdd69d876-dmcfv_9fae323c-a7e9-46a6-b658-1c81c99240e1/keystone-api/0.log" Feb 25 14:17:00 crc kubenswrapper[4815]: I0225 14:17:00.663471 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_1407af2f-82a3-4ccb-9667-03d441ff34dc/kube-state-metrics/0.log" Feb 25 14:17:00 crc kubenswrapper[4815]: I0225 14:17:00.786485 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-85zsm_4e6edebf-da64-4665-9781-b4f3191e0d10/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 25 14:17:01 crc kubenswrapper[4815]: I0225 14:17:01.166289 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-68dffb47c7-mfx8z_2ba9ece6-3c27-41f3-ab94-f4f73b12a265/neutron-httpd/0.log" Feb 25 14:17:01 crc kubenswrapper[4815]: I0225 14:17:01.178862 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-68dffb47c7-mfx8z_2ba9ece6-3c27-41f3-ab94-f4f73b12a265/neutron-api/0.log" Feb 25 14:17:01 crc kubenswrapper[4815]: I0225 14:17:01.387452 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-njs7r_d2b331ae-52bd-4658-ae73-42674c94d3fa/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 25 14:17:01 crc kubenswrapper[4815]: I0225 14:17:01.857469 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_c99a6f77-ea1d-4cb6-848b-447c2ac666c4/nova-cell0-conductor-conductor/0.log" Feb 25 14:17:01 crc kubenswrapper[4815]: I0225 14:17:01.904912 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_f2ffb182-594c-44ce-9d99-32ea22b5bfe9/nova-api-log/0.log" Feb 25 14:17:02 crc kubenswrapper[4815]: I0225 14:17:02.036244 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_f2ffb182-594c-44ce-9d99-32ea22b5bfe9/nova-api-api/0.log" Feb 25 14:17:02 crc kubenswrapper[4815]: I0225 14:17:02.148775 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_3c77e9dd-8e10-48c1-a7ee-7ee4f37a0501/nova-cell1-conductor-conductor/0.log" Feb 25 14:17:02 crc kubenswrapper[4815]: I0225 14:17:02.174947 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_f2597b11-1b7f-4f37-b553-f1385e830db1/nova-cell1-novncproxy-novncproxy/0.log" Feb 25 14:17:02 crc kubenswrapper[4815]: I0225 14:17:02.405680 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-kttx6_84065eef-8fae-4e48-87fa-900a2eab8748/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 25 14:17:02 crc kubenswrapper[4815]: I0225 14:17:02.482292 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d20eee08-f0d7-49bb-a195-202a600195f9/nova-metadata-log/0.log" Feb 25 14:17:02 crc kubenswrapper[4815]: I0225 14:17:02.790881 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_441334b1-15af-4e17-8974-c5ab19338eee/nova-scheduler-scheduler/0.log" Feb 25 14:17:03 crc kubenswrapper[4815]: I0225 14:17:03.029982 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c96edb5d-f9a9-4dc1-a9c2-e3def3efb850/mysql-bootstrap/0.log" Feb 25 14:17:03 crc kubenswrapper[4815]: I0225 14:17:03.207001 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c96edb5d-f9a9-4dc1-a9c2-e3def3efb850/mysql-bootstrap/0.log" Feb 25 14:17:03 crc kubenswrapper[4815]: I0225 14:17:03.226193 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c96edb5d-f9a9-4dc1-a9c2-e3def3efb850/galera/0.log" Feb 25 14:17:03 crc kubenswrapper[4815]: I0225 14:17:03.438201 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_52d1a204-9bb6-40e1-a63c-09ffe6016cd9/mysql-bootstrap/0.log" Feb 25 14:17:03 crc kubenswrapper[4815]: I0225 14:17:03.491494 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d20eee08-f0d7-49bb-a195-202a600195f9/nova-metadata-metadata/0.log" Feb 25 14:17:03 crc kubenswrapper[4815]: I0225 14:17:03.589365 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_52d1a204-9bb6-40e1-a63c-09ffe6016cd9/mysql-bootstrap/0.log" Feb 25 14:17:03 crc kubenswrapper[4815]: I0225 14:17:03.644818 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_52d1a204-9bb6-40e1-a63c-09ffe6016cd9/galera/0.log" Feb 25 14:17:03 crc kubenswrapper[4815]: I0225 14:17:03.686062 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_3f935809-d8cf-4a5f-8d6b-05d7045b6bc8/openstackclient/0.log" Feb 25 14:17:03 crc kubenswrapper[4815]: I0225 14:17:03.923054 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-v2kjv_8e4f98e4-4f35-488a-82e2-a09b2fc2cdfd/openstack-network-exporter/0.log" Feb 25 14:17:03 crc kubenswrapper[4815]: I0225 14:17:03.926594 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-z74w4_67ea29a2-f5eb-41c5-84aa-e1ac90ab567b/ovsdb-server-init/0.log" Feb 25 14:17:04 crc kubenswrapper[4815]: I0225 14:17:04.103467 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-z74w4_67ea29a2-f5eb-41c5-84aa-e1ac90ab567b/ovsdb-server-init/0.log" Feb 25 14:17:04 crc kubenswrapper[4815]: I0225 14:17:04.163339 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-z74w4_67ea29a2-f5eb-41c5-84aa-e1ac90ab567b/ovsdb-server/0.log" Feb 25 14:17:04 crc kubenswrapper[4815]: I0225 14:17:04.226617 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-z74w4_67ea29a2-f5eb-41c5-84aa-e1ac90ab567b/ovs-vswitchd/0.log" Feb 25 14:17:04 crc kubenswrapper[4815]: I0225 14:17:04.385360 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-tjrr6_d9d08116-2a87-4ae4-8485-dab2bb05de4c/ovn-controller/0.log" Feb 25 14:17:04 crc kubenswrapper[4815]: I0225 14:17:04.486991 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-jvq95_8d7f9d84-5806-4beb-9277-c77da4440527/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 25 14:17:04 crc kubenswrapper[4815]: I0225 14:17:04.575923 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ebc28857-40fb-4026-b7df-a2033618af75/openstack-network-exporter/0.log" Feb 25 14:17:04 crc kubenswrapper[4815]: I0225 14:17:04.629777 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ebc28857-40fb-4026-b7df-a2033618af75/ovn-northd/0.log" Feb 25 14:17:04 crc kubenswrapper[4815]: I0225 14:17:04.813591 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_205d03ae-ef40-447e-9176-eb6f6a6d139b/ovsdbserver-nb/0.log" Feb 25 14:17:04 crc kubenswrapper[4815]: I0225 14:17:04.837238 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_205d03ae-ef40-447e-9176-eb6f6a6d139b/openstack-network-exporter/0.log" Feb 25 14:17:04 crc kubenswrapper[4815]: I0225 14:17:04.950011 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_16d450f1-ecbb-4266-9818-afb1470b3a71/openstack-network-exporter/0.log" Feb 25 14:17:05 crc kubenswrapper[4815]: I0225 14:17:05.011223 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_16d450f1-ecbb-4266-9818-afb1470b3a71/ovsdbserver-sb/0.log" Feb 25 14:17:05 crc kubenswrapper[4815]: I0225 14:17:05.159719 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-b695645c6-vszf4_99ea30bb-39c5-47ee-aa2a-008fe013de18/placement-api/0.log" Feb 25 14:17:05 crc kubenswrapper[4815]: I0225 14:17:05.216887 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-b695645c6-vszf4_99ea30bb-39c5-47ee-aa2a-008fe013de18/placement-log/0.log" Feb 25 14:17:05 crc kubenswrapper[4815]: I0225 14:17:05.335258 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1d3e5784-65f2-4282-a148-b4b015e5f875/setup-container/0.log" Feb 25 14:17:05 crc kubenswrapper[4815]: I0225 14:17:05.479344 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1d3e5784-65f2-4282-a148-b4b015e5f875/setup-container/0.log" Feb 25 14:17:05 crc kubenswrapper[4815]: I0225 14:17:05.585744 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1d3e5784-65f2-4282-a148-b4b015e5f875/rabbitmq/0.log" Feb 25 14:17:05 crc kubenswrapper[4815]: I0225 14:17:05.605530 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bd5b84a1-88ec-4727-9683-e696540ec165/setup-container/0.log" Feb 25 14:17:05 crc kubenswrapper[4815]: I0225 14:17:05.822891 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bd5b84a1-88ec-4727-9683-e696540ec165/setup-container/0.log" Feb 25 14:17:05 crc kubenswrapper[4815]: I0225 14:17:05.828616 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bd5b84a1-88ec-4727-9683-e696540ec165/rabbitmq/0.log" Feb 25 14:17:05 crc kubenswrapper[4815]: I0225 14:17:05.849741 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-tn866_55562f1f-6c46-4413-9f01-99867d3a1a7f/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 25 14:17:06 crc kubenswrapper[4815]: I0225 14:17:06.021903 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-t8psr_017b7720-a334-4276-9531-a49fb9558275/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 25 14:17:06 crc kubenswrapper[4815]: I0225 14:17:06.127410 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-b6k6j_5bcbebb7-6865-44e4-b4da-61c06daeb9de/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 25 14:17:06 crc kubenswrapper[4815]: I0225 14:17:06.356067 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-xkxrf_c037c85e-2d64-428a-a7c7-bcdabbb57676/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 25 14:17:06 crc kubenswrapper[4815]: I0225 14:17:06.409901 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-lfl6v_80323484-d533-4ca8-bfa5-315d318053f5/ssh-known-hosts-edpm-deployment/0.log" Feb 25 14:17:06 crc kubenswrapper[4815]: I0225 14:17:06.629073 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-68d8cd5787-kmnfm_ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96/proxy-server/0.log" Feb 25 14:17:06 crc kubenswrapper[4815]: I0225 14:17:06.669155 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-68d8cd5787-kmnfm_ecf8b0bf-4649-4cfd-9e3a-64d7f3fc4d96/proxy-httpd/0.log" Feb 25 14:17:06 crc kubenswrapper[4815]: I0225 14:17:06.801295 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-zplf2_3ade0feb-5e78-4cf2-a82e-bad0e5468bc7/swift-ring-rebalance/0.log" Feb 25 14:17:06 crc kubenswrapper[4815]: I0225 14:17:06.890309 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3941650f-c99b-4043-9137-e4c43361f93f/account-auditor/0.log" Feb 25 14:17:06 crc kubenswrapper[4815]: I0225 14:17:06.893480 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3941650f-c99b-4043-9137-e4c43361f93f/account-reaper/0.log" Feb 25 14:17:07 crc kubenswrapper[4815]: I0225 14:17:07.035934 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3941650f-c99b-4043-9137-e4c43361f93f/account-replicator/0.log" Feb 25 14:17:07 crc kubenswrapper[4815]: I0225 14:17:07.090911 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3941650f-c99b-4043-9137-e4c43361f93f/account-server/0.log" Feb 25 14:17:07 crc kubenswrapper[4815]: I0225 14:17:07.165882 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3941650f-c99b-4043-9137-e4c43361f93f/container-auditor/0.log" Feb 25 14:17:07 crc kubenswrapper[4815]: I0225 14:17:07.225323 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3941650f-c99b-4043-9137-e4c43361f93f/container-replicator/0.log" Feb 25 14:17:07 crc kubenswrapper[4815]: I0225 14:17:07.253001 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3941650f-c99b-4043-9137-e4c43361f93f/container-server/0.log" Feb 25 14:17:07 crc kubenswrapper[4815]: I0225 14:17:07.306380 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3941650f-c99b-4043-9137-e4c43361f93f/container-updater/0.log" Feb 25 14:17:07 crc kubenswrapper[4815]: I0225 14:17:07.372180 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3941650f-c99b-4043-9137-e4c43361f93f/object-auditor/0.log" Feb 25 14:17:07 crc kubenswrapper[4815]: I0225 14:17:07.453519 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3941650f-c99b-4043-9137-e4c43361f93f/object-expirer/0.log" Feb 25 14:17:07 crc kubenswrapper[4815]: I0225 14:17:07.502547 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3941650f-c99b-4043-9137-e4c43361f93f/object-replicator/0.log" Feb 25 14:17:07 crc kubenswrapper[4815]: I0225 14:17:07.522133 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3941650f-c99b-4043-9137-e4c43361f93f/object-server/0.log" Feb 25 14:17:07 crc kubenswrapper[4815]: I0225 14:17:07.654439 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3941650f-c99b-4043-9137-e4c43361f93f/object-updater/0.log" Feb 25 14:17:07 crc kubenswrapper[4815]: I0225 14:17:07.709784 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3941650f-c99b-4043-9137-e4c43361f93f/rsync/0.log" Feb 25 14:17:07 crc kubenswrapper[4815]: I0225 14:17:07.715966 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_3941650f-c99b-4043-9137-e4c43361f93f/swift-recon-cron/0.log" Feb 25 14:17:07 crc kubenswrapper[4815]: I0225 14:17:07.937583 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-47nxh_1b1ea76e-0503-4e5f-8adb-ad5e39e5bea6/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 25 14:17:08 crc kubenswrapper[4815]: I0225 14:17:08.106381 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_d23e8f04-361c-4227-a5c9-bfcd4801f771/tempest-tests-tempest-tests-runner/0.log" Feb 25 14:17:08 crc kubenswrapper[4815]: I0225 14:17:08.137814 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_91988324-9ed0-4293-82e5-738a92f680e1/test-operator-logs-container/0.log" Feb 25 14:17:08 crc kubenswrapper[4815]: I0225 14:17:08.264220 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-xf7vb_9fdfaae5-e949-48ed-b3cf-a8893dc84a67/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 25 14:17:12 crc kubenswrapper[4815]: I0225 14:17:12.989408 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_90528e4b-0537-4477-b7d1-603c21bc9358/memcached/0.log" Feb 25 14:17:21 crc kubenswrapper[4815]: I0225 14:17:21.341882 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 14:17:21 crc kubenswrapper[4815]: I0225 14:17:21.342437 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 14:17:31 crc kubenswrapper[4815]: I0225 14:17:31.875636 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d_01da6351-8eb8-427f-aaa2-73d2a7bcaac6/util/0.log" Feb 25 14:17:32 crc kubenswrapper[4815]: I0225 14:17:32.100276 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d_01da6351-8eb8-427f-aaa2-73d2a7bcaac6/pull/0.log" Feb 25 14:17:32 crc kubenswrapper[4815]: I0225 14:17:32.118182 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d_01da6351-8eb8-427f-aaa2-73d2a7bcaac6/pull/0.log" Feb 25 14:17:32 crc kubenswrapper[4815]: I0225 14:17:32.135545 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d_01da6351-8eb8-427f-aaa2-73d2a7bcaac6/util/0.log" Feb 25 14:17:32 crc kubenswrapper[4815]: I0225 14:17:32.317385 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d_01da6351-8eb8-427f-aaa2-73d2a7bcaac6/util/0.log" Feb 25 14:17:32 crc kubenswrapper[4815]: I0225 14:17:32.328166 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d_01da6351-8eb8-427f-aaa2-73d2a7bcaac6/extract/0.log" Feb 25 14:17:32 crc kubenswrapper[4815]: I0225 14:17:32.347342 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_37c49f77305b81cfdbd654b856809a15d63e7f104126676e449bca903c8x84d_01da6351-8eb8-427f-aaa2-73d2a7bcaac6/pull/0.log" Feb 25 14:17:32 crc kubenswrapper[4815]: I0225 14:17:32.756620 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-qm5sw_c4956173-674d-4b10-9a3f-9996f9190299/manager/0.log" Feb 25 14:17:33 crc kubenswrapper[4815]: I0225 14:17:33.117978 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-784b5bb6c5-gw8z5_9fa8caf9-fab3-44d2-9841-593558e9a690/manager/0.log" Feb 25 14:17:33 crc kubenswrapper[4815]: I0225 14:17:33.228700 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-csfk2_50963101-0a9e-453d-925b-ac6f08c545c4/manager/0.log" Feb 25 14:17:33 crc kubenswrapper[4815]: I0225 14:17:33.517470 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-ccl5w_33e59af6-b9bb-4b55-b6a9-c6963e269b94/manager/0.log" Feb 25 14:17:33 crc kubenswrapper[4815]: I0225 14:17:33.900971 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-55d77d7b5c-h9n4d_bace17a8-ce7d-4df9-8259-68931eeecde8/manager/0.log" Feb 25 14:17:33 crc kubenswrapper[4815]: I0225 14:17:33.935638 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-52psd_12b1d896-ab7e-4097-afa4-791d39e1924e/manager/0.log" Feb 25 14:17:34 crc kubenswrapper[4815]: I0225 14:17:34.156960 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79d975b745-s5ztn_9a554278-1139-4942-904a-ffc60adde6de/manager/0.log" Feb 25 14:17:34 crc kubenswrapper[4815]: I0225 14:17:34.404990 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-rx4fg_c64a07be-3cda-473e-89b6-fd0347519c93/manager/0.log" Feb 25 14:17:34 crc kubenswrapper[4815]: I0225 14:17:34.595181 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-67d996989d-5fq96_4dc8d07d-3e17-40a8-ac19-187729a64c4d/manager/0.log" Feb 25 14:17:34 crc kubenswrapper[4815]: I0225 14:17:34.658255 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-8t7p2_16cb48d6-2014-44ec-aaf5-6aef86d24cdf/manager/0.log" Feb 25 14:17:34 crc kubenswrapper[4815]: I0225 14:17:34.889025 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6bd4687957-vnb57_6bab4cd6-5eff-409c-83e3-ad5c1a6b6444/manager/0.log" Feb 25 14:17:35 crc kubenswrapper[4815]: I0225 14:17:35.191814 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-659dc6bbfc-qtr8c_b7cddc29-1c58-4edd-992f-165386b751ba/manager/0.log" Feb 25 14:17:35 crc kubenswrapper[4815]: I0225 14:17:35.205932 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-z4jrr_b30720e2-30f0-480e-a44f-2b409be766c6/manager/0.log" Feb 25 14:17:35 crc kubenswrapper[4815]: I0225 14:17:35.403192 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c6767dc9c7vg7p_43faa7ed-99a7-478c-8994-7a97333df231/manager/0.log" Feb 25 14:17:35 crc kubenswrapper[4815]: I0225 14:17:35.775290 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-c7f69bdc4-8vjkl_a1251cc6-4b50-4551-b051-028595a75965/operator/0.log" Feb 25 14:17:36 crc kubenswrapper[4815]: I0225 14:17:36.000174 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-fgf9z_9a2ffc78-6a9a-45b2-81ab-bfe65898a0cd/registry-server/0.log" Feb 25 14:17:36 crc kubenswrapper[4815]: I0225 14:17:36.328975 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5955d8c787-62tw8_697c602f-d6bc-478c-b89b-e3cc8aee4dcc/manager/0.log" Feb 25 14:17:36 crc kubenswrapper[4815]: I0225 14:17:36.496075 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-9qrdg_3b1a8f42-91a7-46b3-96c6-0ef20a126678/manager/0.log" Feb 25 14:17:36 crc kubenswrapper[4815]: I0225 14:17:36.756730 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-7l95p_22b232e5-5045-404a-a248-01babe4886c3/operator/0.log" Feb 25 14:17:36 crc kubenswrapper[4815]: I0225 14:17:36.851871 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-8vr5p_0673a594-2080-4325-80ce-f5597ad337c7/manager/0.log" Feb 25 14:17:37 crc kubenswrapper[4815]: I0225 14:17:37.065278 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-589c568786-rwzjh_136c32bc-0a85-4a36-b527-a787fa7158f1/manager/0.log" Feb 25 14:17:37 crc kubenswrapper[4815]: I0225 14:17:37.106748 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5dc6794d5b-grtsf_21de3543-3d6a-4adf-b104-f9fcaee4532b/manager/0.log" Feb 25 14:17:37 crc kubenswrapper[4815]: I0225 14:17:37.314235 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-bccc79885-v825g_486a2442-b575-491d-9c18-4185e5a647fe/manager/0.log" Feb 25 14:17:37 crc kubenswrapper[4815]: I0225 14:17:37.411451 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5f47dfdb6c-slw4h_509894d6-7814-4f85-9eed-f1946b029dbf/manager/0.log" Feb 25 14:17:38 crc kubenswrapper[4815]: I0225 14:17:38.882183 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-zwknl_62c00caf-05cf-4229-9349-341a2b597b12/manager/0.log" Feb 25 14:17:51 crc kubenswrapper[4815]: I0225 14:17:51.341961 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 14:17:51 crc kubenswrapper[4815]: I0225 14:17:51.342498 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 14:17:51 crc kubenswrapper[4815]: I0225 14:17:51.342556 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 14:17:51 crc kubenswrapper[4815]: I0225 14:17:51.343295 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"35ad0ed6ac05ca7660eb380c1f3eb06169b51f834d4f9b66e462803955bc5ff9"} pod="openshift-machine-config-operator/machine-config-daemon-frc75" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 14:17:51 crc kubenswrapper[4815]: I0225 14:17:51.343356 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" containerID="cri-o://35ad0ed6ac05ca7660eb380c1f3eb06169b51f834d4f9b66e462803955bc5ff9" gracePeriod=600 Feb 25 14:17:51 crc kubenswrapper[4815]: I0225 14:17:51.791770 4815 generic.go:334] "Generic (PLEG): container finished" podID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerID="35ad0ed6ac05ca7660eb380c1f3eb06169b51f834d4f9b66e462803955bc5ff9" exitCode=0 Feb 25 14:17:51 crc kubenswrapper[4815]: I0225 14:17:51.791850 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerDied","Data":"35ad0ed6ac05ca7660eb380c1f3eb06169b51f834d4f9b66e462803955bc5ff9"} Feb 25 14:17:51 crc kubenswrapper[4815]: I0225 14:17:51.792098 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerStarted","Data":"8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107"} Feb 25 14:17:51 crc kubenswrapper[4815]: I0225 14:17:51.792128 4815 scope.go:117] "RemoveContainer" containerID="68f7e9b6478bb0d949df6eab39d39a57d4ddc4932629604ba5eaec7d84a7c362" Feb 25 14:17:54 crc kubenswrapper[4815]: I0225 14:17:54.614645 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8ltgt"] Feb 25 14:17:54 crc kubenswrapper[4815]: E0225 14:17:54.615441 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c53cc96d-fb25-4a3e-874d-731604187c5a" containerName="container-00" Feb 25 14:17:54 crc kubenswrapper[4815]: I0225 14:17:54.615452 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="c53cc96d-fb25-4a3e-874d-731604187c5a" containerName="container-00" Feb 25 14:17:54 crc kubenswrapper[4815]: I0225 14:17:54.615652 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="c53cc96d-fb25-4a3e-874d-731604187c5a" containerName="container-00" Feb 25 14:17:54 crc kubenswrapper[4815]: I0225 14:17:54.616908 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8ltgt" Feb 25 14:17:54 crc kubenswrapper[4815]: I0225 14:17:54.626305 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8ltgt"] Feb 25 14:17:54 crc kubenswrapper[4815]: I0225 14:17:54.767681 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c1c12e-5e32-440c-a6fd-c33486abd377-catalog-content\") pod \"redhat-operators-8ltgt\" (UID: \"82c1c12e-5e32-440c-a6fd-c33486abd377\") " pod="openshift-marketplace/redhat-operators-8ltgt" Feb 25 14:17:54 crc kubenswrapper[4815]: I0225 14:17:54.767731 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c1c12e-5e32-440c-a6fd-c33486abd377-utilities\") pod \"redhat-operators-8ltgt\" (UID: \"82c1c12e-5e32-440c-a6fd-c33486abd377\") " pod="openshift-marketplace/redhat-operators-8ltgt" Feb 25 14:17:54 crc kubenswrapper[4815]: I0225 14:17:54.767784 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27n9k\" (UniqueName: \"kubernetes.io/projected/82c1c12e-5e32-440c-a6fd-c33486abd377-kube-api-access-27n9k\") pod \"redhat-operators-8ltgt\" (UID: \"82c1c12e-5e32-440c-a6fd-c33486abd377\") " pod="openshift-marketplace/redhat-operators-8ltgt" Feb 25 14:17:54 crc kubenswrapper[4815]: I0225 14:17:54.869815 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c1c12e-5e32-440c-a6fd-c33486abd377-catalog-content\") pod \"redhat-operators-8ltgt\" (UID: \"82c1c12e-5e32-440c-a6fd-c33486abd377\") " pod="openshift-marketplace/redhat-operators-8ltgt" Feb 25 14:17:54 crc kubenswrapper[4815]: I0225 14:17:54.869889 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c1c12e-5e32-440c-a6fd-c33486abd377-utilities\") pod \"redhat-operators-8ltgt\" (UID: \"82c1c12e-5e32-440c-a6fd-c33486abd377\") " pod="openshift-marketplace/redhat-operators-8ltgt" Feb 25 14:17:54 crc kubenswrapper[4815]: I0225 14:17:54.869947 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27n9k\" (UniqueName: \"kubernetes.io/projected/82c1c12e-5e32-440c-a6fd-c33486abd377-kube-api-access-27n9k\") pod \"redhat-operators-8ltgt\" (UID: \"82c1c12e-5e32-440c-a6fd-c33486abd377\") " pod="openshift-marketplace/redhat-operators-8ltgt" Feb 25 14:17:54 crc kubenswrapper[4815]: I0225 14:17:54.870317 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c1c12e-5e32-440c-a6fd-c33486abd377-catalog-content\") pod \"redhat-operators-8ltgt\" (UID: \"82c1c12e-5e32-440c-a6fd-c33486abd377\") " pod="openshift-marketplace/redhat-operators-8ltgt" Feb 25 14:17:54 crc kubenswrapper[4815]: I0225 14:17:54.870352 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c1c12e-5e32-440c-a6fd-c33486abd377-utilities\") pod \"redhat-operators-8ltgt\" (UID: \"82c1c12e-5e32-440c-a6fd-c33486abd377\") " pod="openshift-marketplace/redhat-operators-8ltgt" Feb 25 14:17:54 crc kubenswrapper[4815]: I0225 14:17:54.893908 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27n9k\" (UniqueName: \"kubernetes.io/projected/82c1c12e-5e32-440c-a6fd-c33486abd377-kube-api-access-27n9k\") pod \"redhat-operators-8ltgt\" (UID: \"82c1c12e-5e32-440c-a6fd-c33486abd377\") " pod="openshift-marketplace/redhat-operators-8ltgt" Feb 25 14:17:54 crc kubenswrapper[4815]: I0225 14:17:54.934051 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8ltgt" Feb 25 14:17:55 crc kubenswrapper[4815]: I0225 14:17:55.426100 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8ltgt"] Feb 25 14:17:55 crc kubenswrapper[4815]: W0225 14:17:55.428200 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82c1c12e_5e32_440c_a6fd_c33486abd377.slice/crio-8ea02de7ee41da11031cf1de187f883a914d42aa506fe05adbf9565216d63d4f WatchSource:0}: Error finding container 8ea02de7ee41da11031cf1de187f883a914d42aa506fe05adbf9565216d63d4f: Status 404 returned error can't find the container with id 8ea02de7ee41da11031cf1de187f883a914d42aa506fe05adbf9565216d63d4f Feb 25 14:17:55 crc kubenswrapper[4815]: I0225 14:17:55.832168 4815 generic.go:334] "Generic (PLEG): container finished" podID="82c1c12e-5e32-440c-a6fd-c33486abd377" containerID="2a81fec41e74683784c79d39df61a55233b276093ed85424a0e149fafb244f27" exitCode=0 Feb 25 14:17:55 crc kubenswrapper[4815]: I0225 14:17:55.832475 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ltgt" event={"ID":"82c1c12e-5e32-440c-a6fd-c33486abd377","Type":"ContainerDied","Data":"2a81fec41e74683784c79d39df61a55233b276093ed85424a0e149fafb244f27"} Feb 25 14:17:55 crc kubenswrapper[4815]: I0225 14:17:55.832580 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ltgt" event={"ID":"82c1c12e-5e32-440c-a6fd-c33486abd377","Type":"ContainerStarted","Data":"8ea02de7ee41da11031cf1de187f883a914d42aa506fe05adbf9565216d63d4f"} Feb 25 14:17:55 crc kubenswrapper[4815]: I0225 14:17:55.835546 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 14:17:56 crc kubenswrapper[4815]: I0225 14:17:56.843135 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ltgt" event={"ID":"82c1c12e-5e32-440c-a6fd-c33486abd377","Type":"ContainerStarted","Data":"edd164af69ad35f8010d8b77028750c213aea50bf26025ccdb2365a991f8f630"} Feb 25 14:17:57 crc kubenswrapper[4815]: I0225 14:17:57.853419 4815 generic.go:334] "Generic (PLEG): container finished" podID="82c1c12e-5e32-440c-a6fd-c33486abd377" containerID="edd164af69ad35f8010d8b77028750c213aea50bf26025ccdb2365a991f8f630" exitCode=0 Feb 25 14:17:57 crc kubenswrapper[4815]: I0225 14:17:57.853586 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ltgt" event={"ID":"82c1c12e-5e32-440c-a6fd-c33486abd377","Type":"ContainerDied","Data":"edd164af69ad35f8010d8b77028750c213aea50bf26025ccdb2365a991f8f630"} Feb 25 14:17:58 crc kubenswrapper[4815]: I0225 14:17:58.422440 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-x5wvt_61c629d8-0773-44ce-8919-b32d32aa5d11/control-plane-machine-set-operator/0.log" Feb 25 14:17:58 crc kubenswrapper[4815]: I0225 14:17:58.661623 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6glz5_6f28beca-7506-4390-9c74-86466a6b09ef/kube-rbac-proxy/0.log" Feb 25 14:17:58 crc kubenswrapper[4815]: I0225 14:17:58.697917 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6glz5_6f28beca-7506-4390-9c74-86466a6b09ef/machine-api-operator/0.log" Feb 25 14:17:58 crc kubenswrapper[4815]: I0225 14:17:58.866338 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ltgt" event={"ID":"82c1c12e-5e32-440c-a6fd-c33486abd377","Type":"ContainerStarted","Data":"0f7f479e622fc30d63a6f400bd1acc39379439cd86930ec84666ab3a6afd9ba5"} Feb 25 14:18:00 crc kubenswrapper[4815]: I0225 14:18:00.144792 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8ltgt" podStartSLOduration=3.6858987279999997 podStartE2EDuration="6.144772498s" podCreationTimestamp="2026-02-25 14:17:54 +0000 UTC" firstStartedPulling="2026-02-25 14:17:55.835242137 +0000 UTC m=+3433.636340191" lastFinishedPulling="2026-02-25 14:17:58.294115907 +0000 UTC m=+3436.095213961" observedRunningTime="2026-02-25 14:17:58.891948521 +0000 UTC m=+3436.693046585" watchObservedRunningTime="2026-02-25 14:18:00.144772498 +0000 UTC m=+3437.945870552" Feb 25 14:18:00 crc kubenswrapper[4815]: I0225 14:18:00.147092 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533818-2wncd"] Feb 25 14:18:00 crc kubenswrapper[4815]: I0225 14:18:00.148193 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533818-2wncd" Feb 25 14:18:00 crc kubenswrapper[4815]: I0225 14:18:00.172385 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 14:18:00 crc kubenswrapper[4815]: I0225 14:18:00.172607 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 14:18:00 crc kubenswrapper[4815]: I0225 14:18:00.172791 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 14:18:00 crc kubenswrapper[4815]: I0225 14:18:00.181408 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533818-2wncd"] Feb 25 14:18:00 crc kubenswrapper[4815]: I0225 14:18:00.272579 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d4nz\" (UniqueName: \"kubernetes.io/projected/23253ded-9aae-451f-a7fc-1b10518f4312-kube-api-access-5d4nz\") pod \"auto-csr-approver-29533818-2wncd\" (UID: \"23253ded-9aae-451f-a7fc-1b10518f4312\") " pod="openshift-infra/auto-csr-approver-29533818-2wncd" Feb 25 14:18:00 crc kubenswrapper[4815]: I0225 14:18:00.374527 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d4nz\" (UniqueName: \"kubernetes.io/projected/23253ded-9aae-451f-a7fc-1b10518f4312-kube-api-access-5d4nz\") pod \"auto-csr-approver-29533818-2wncd\" (UID: \"23253ded-9aae-451f-a7fc-1b10518f4312\") " pod="openshift-infra/auto-csr-approver-29533818-2wncd" Feb 25 14:18:00 crc kubenswrapper[4815]: I0225 14:18:00.397600 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d4nz\" (UniqueName: \"kubernetes.io/projected/23253ded-9aae-451f-a7fc-1b10518f4312-kube-api-access-5d4nz\") pod \"auto-csr-approver-29533818-2wncd\" (UID: \"23253ded-9aae-451f-a7fc-1b10518f4312\") " pod="openshift-infra/auto-csr-approver-29533818-2wncd" Feb 25 14:18:00 crc kubenswrapper[4815]: I0225 14:18:00.493033 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533818-2wncd" Feb 25 14:18:00 crc kubenswrapper[4815]: I0225 14:18:00.976549 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533818-2wncd"] Feb 25 14:18:01 crc kubenswrapper[4815]: I0225 14:18:01.890383 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533818-2wncd" event={"ID":"23253ded-9aae-451f-a7fc-1b10518f4312","Type":"ContainerStarted","Data":"756a765217522e90a819a3d45cd65977dd84a9d02c8ad72bf0ac690d989307e4"} Feb 25 14:18:02 crc kubenswrapper[4815]: I0225 14:18:02.900681 4815 generic.go:334] "Generic (PLEG): container finished" podID="23253ded-9aae-451f-a7fc-1b10518f4312" containerID="6b18ca2315823c26b460d836ee0e1a4578448d2af942774f6b847a88525d1622" exitCode=0 Feb 25 14:18:02 crc kubenswrapper[4815]: I0225 14:18:02.900753 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533818-2wncd" event={"ID":"23253ded-9aae-451f-a7fc-1b10518f4312","Type":"ContainerDied","Data":"6b18ca2315823c26b460d836ee0e1a4578448d2af942774f6b847a88525d1622"} Feb 25 14:18:04 crc kubenswrapper[4815]: I0225 14:18:04.272451 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533818-2wncd" Feb 25 14:18:04 crc kubenswrapper[4815]: I0225 14:18:04.473533 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5d4nz\" (UniqueName: \"kubernetes.io/projected/23253ded-9aae-451f-a7fc-1b10518f4312-kube-api-access-5d4nz\") pod \"23253ded-9aae-451f-a7fc-1b10518f4312\" (UID: \"23253ded-9aae-451f-a7fc-1b10518f4312\") " Feb 25 14:18:04 crc kubenswrapper[4815]: I0225 14:18:04.483953 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23253ded-9aae-451f-a7fc-1b10518f4312-kube-api-access-5d4nz" (OuterVolumeSpecName: "kube-api-access-5d4nz") pod "23253ded-9aae-451f-a7fc-1b10518f4312" (UID: "23253ded-9aae-451f-a7fc-1b10518f4312"). InnerVolumeSpecName "kube-api-access-5d4nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:18:04 crc kubenswrapper[4815]: I0225 14:18:04.575039 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5d4nz\" (UniqueName: \"kubernetes.io/projected/23253ded-9aae-451f-a7fc-1b10518f4312-kube-api-access-5d4nz\") on node \"crc\" DevicePath \"\"" Feb 25 14:18:04 crc kubenswrapper[4815]: I0225 14:18:04.918970 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533818-2wncd" event={"ID":"23253ded-9aae-451f-a7fc-1b10518f4312","Type":"ContainerDied","Data":"756a765217522e90a819a3d45cd65977dd84a9d02c8ad72bf0ac690d989307e4"} Feb 25 14:18:04 crc kubenswrapper[4815]: I0225 14:18:04.919030 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="756a765217522e90a819a3d45cd65977dd84a9d02c8ad72bf0ac690d989307e4" Feb 25 14:18:04 crc kubenswrapper[4815]: I0225 14:18:04.919093 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533818-2wncd" Feb 25 14:18:04 crc kubenswrapper[4815]: I0225 14:18:04.934177 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8ltgt" Feb 25 14:18:04 crc kubenswrapper[4815]: I0225 14:18:04.954438 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8ltgt" Feb 25 14:18:04 crc kubenswrapper[4815]: I0225 14:18:04.998876 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8ltgt" Feb 25 14:18:05 crc kubenswrapper[4815]: I0225 14:18:05.345713 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533812-52hp2"] Feb 25 14:18:05 crc kubenswrapper[4815]: I0225 14:18:05.353324 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533812-52hp2"] Feb 25 14:18:05 crc kubenswrapper[4815]: I0225 14:18:05.979323 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8ltgt" Feb 25 14:18:06 crc kubenswrapper[4815]: I0225 14:18:06.041940 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8ltgt"] Feb 25 14:18:06 crc kubenswrapper[4815]: I0225 14:18:06.949781 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78613b37-9fbb-4ef7-8c1c-660da24efdbe" path="/var/lib/kubelet/pods/78613b37-9fbb-4ef7-8c1c-660da24efdbe/volumes" Feb 25 14:18:07 crc kubenswrapper[4815]: I0225 14:18:07.944180 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8ltgt" podUID="82c1c12e-5e32-440c-a6fd-c33486abd377" containerName="registry-server" containerID="cri-o://0f7f479e622fc30d63a6f400bd1acc39379439cd86930ec84666ab3a6afd9ba5" gracePeriod=2 Feb 25 14:18:08 crc kubenswrapper[4815]: E0225 14:18:08.171044 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82c1c12e_5e32_440c_a6fd_c33486abd377.slice/crio-0f7f479e622fc30d63a6f400bd1acc39379439cd86930ec84666ab3a6afd9ba5.scope\": RecentStats: unable to find data in memory cache]" Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.395771 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8ltgt" Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.546034 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c1c12e-5e32-440c-a6fd-c33486abd377-utilities\") pod \"82c1c12e-5e32-440c-a6fd-c33486abd377\" (UID: \"82c1c12e-5e32-440c-a6fd-c33486abd377\") " Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.546087 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c1c12e-5e32-440c-a6fd-c33486abd377-catalog-content\") pod \"82c1c12e-5e32-440c-a6fd-c33486abd377\" (UID: \"82c1c12e-5e32-440c-a6fd-c33486abd377\") " Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.546143 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27n9k\" (UniqueName: \"kubernetes.io/projected/82c1c12e-5e32-440c-a6fd-c33486abd377-kube-api-access-27n9k\") pod \"82c1c12e-5e32-440c-a6fd-c33486abd377\" (UID: \"82c1c12e-5e32-440c-a6fd-c33486abd377\") " Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.546915 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82c1c12e-5e32-440c-a6fd-c33486abd377-utilities" (OuterVolumeSpecName: "utilities") pod "82c1c12e-5e32-440c-a6fd-c33486abd377" (UID: "82c1c12e-5e32-440c-a6fd-c33486abd377"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.564941 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82c1c12e-5e32-440c-a6fd-c33486abd377-kube-api-access-27n9k" (OuterVolumeSpecName: "kube-api-access-27n9k") pod "82c1c12e-5e32-440c-a6fd-c33486abd377" (UID: "82c1c12e-5e32-440c-a6fd-c33486abd377"). InnerVolumeSpecName "kube-api-access-27n9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.648838 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c1c12e-5e32-440c-a6fd-c33486abd377-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.648866 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27n9k\" (UniqueName: \"kubernetes.io/projected/82c1c12e-5e32-440c-a6fd-c33486abd377-kube-api-access-27n9k\") on node \"crc\" DevicePath \"\"" Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.669343 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82c1c12e-5e32-440c-a6fd-c33486abd377-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82c1c12e-5e32-440c-a6fd-c33486abd377" (UID: "82c1c12e-5e32-440c-a6fd-c33486abd377"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.751169 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c1c12e-5e32-440c-a6fd-c33486abd377-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.957528 4815 generic.go:334] "Generic (PLEG): container finished" podID="82c1c12e-5e32-440c-a6fd-c33486abd377" containerID="0f7f479e622fc30d63a6f400bd1acc39379439cd86930ec84666ab3a6afd9ba5" exitCode=0 Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.957577 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ltgt" event={"ID":"82c1c12e-5e32-440c-a6fd-c33486abd377","Type":"ContainerDied","Data":"0f7f479e622fc30d63a6f400bd1acc39379439cd86930ec84666ab3a6afd9ba5"} Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.957610 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ltgt" event={"ID":"82c1c12e-5e32-440c-a6fd-c33486abd377","Type":"ContainerDied","Data":"8ea02de7ee41da11031cf1de187f883a914d42aa506fe05adbf9565216d63d4f"} Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.957636 4815 scope.go:117] "RemoveContainer" containerID="0f7f479e622fc30d63a6f400bd1acc39379439cd86930ec84666ab3a6afd9ba5" Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.957626 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8ltgt" Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.982218 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8ltgt"] Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.987333 4815 scope.go:117] "RemoveContainer" containerID="edd164af69ad35f8010d8b77028750c213aea50bf26025ccdb2365a991f8f630" Feb 25 14:18:08 crc kubenswrapper[4815]: I0225 14:18:08.989608 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8ltgt"] Feb 25 14:18:09 crc kubenswrapper[4815]: I0225 14:18:09.010374 4815 scope.go:117] "RemoveContainer" containerID="2a81fec41e74683784c79d39df61a55233b276093ed85424a0e149fafb244f27" Feb 25 14:18:09 crc kubenswrapper[4815]: I0225 14:18:09.057001 4815 scope.go:117] "RemoveContainer" containerID="0f7f479e622fc30d63a6f400bd1acc39379439cd86930ec84666ab3a6afd9ba5" Feb 25 14:18:09 crc kubenswrapper[4815]: E0225 14:18:09.057925 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f7f479e622fc30d63a6f400bd1acc39379439cd86930ec84666ab3a6afd9ba5\": container with ID starting with 0f7f479e622fc30d63a6f400bd1acc39379439cd86930ec84666ab3a6afd9ba5 not found: ID does not exist" containerID="0f7f479e622fc30d63a6f400bd1acc39379439cd86930ec84666ab3a6afd9ba5" Feb 25 14:18:09 crc kubenswrapper[4815]: I0225 14:18:09.057973 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f7f479e622fc30d63a6f400bd1acc39379439cd86930ec84666ab3a6afd9ba5"} err="failed to get container status \"0f7f479e622fc30d63a6f400bd1acc39379439cd86930ec84666ab3a6afd9ba5\": rpc error: code = NotFound desc = could not find container \"0f7f479e622fc30d63a6f400bd1acc39379439cd86930ec84666ab3a6afd9ba5\": container with ID starting with 0f7f479e622fc30d63a6f400bd1acc39379439cd86930ec84666ab3a6afd9ba5 not found: ID does not exist" Feb 25 14:18:09 crc kubenswrapper[4815]: I0225 14:18:09.058004 4815 scope.go:117] "RemoveContainer" containerID="edd164af69ad35f8010d8b77028750c213aea50bf26025ccdb2365a991f8f630" Feb 25 14:18:09 crc kubenswrapper[4815]: E0225 14:18:09.058292 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edd164af69ad35f8010d8b77028750c213aea50bf26025ccdb2365a991f8f630\": container with ID starting with edd164af69ad35f8010d8b77028750c213aea50bf26025ccdb2365a991f8f630 not found: ID does not exist" containerID="edd164af69ad35f8010d8b77028750c213aea50bf26025ccdb2365a991f8f630" Feb 25 14:18:09 crc kubenswrapper[4815]: I0225 14:18:09.058319 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edd164af69ad35f8010d8b77028750c213aea50bf26025ccdb2365a991f8f630"} err="failed to get container status \"edd164af69ad35f8010d8b77028750c213aea50bf26025ccdb2365a991f8f630\": rpc error: code = NotFound desc = could not find container \"edd164af69ad35f8010d8b77028750c213aea50bf26025ccdb2365a991f8f630\": container with ID starting with edd164af69ad35f8010d8b77028750c213aea50bf26025ccdb2365a991f8f630 not found: ID does not exist" Feb 25 14:18:09 crc kubenswrapper[4815]: I0225 14:18:09.058342 4815 scope.go:117] "RemoveContainer" containerID="2a81fec41e74683784c79d39df61a55233b276093ed85424a0e149fafb244f27" Feb 25 14:18:09 crc kubenswrapper[4815]: E0225 14:18:09.058566 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a81fec41e74683784c79d39df61a55233b276093ed85424a0e149fafb244f27\": container with ID starting with 2a81fec41e74683784c79d39df61a55233b276093ed85424a0e149fafb244f27 not found: ID does not exist" containerID="2a81fec41e74683784c79d39df61a55233b276093ed85424a0e149fafb244f27" Feb 25 14:18:09 crc kubenswrapper[4815]: I0225 14:18:09.058597 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a81fec41e74683784c79d39df61a55233b276093ed85424a0e149fafb244f27"} err="failed to get container status \"2a81fec41e74683784c79d39df61a55233b276093ed85424a0e149fafb244f27\": rpc error: code = NotFound desc = could not find container \"2a81fec41e74683784c79d39df61a55233b276093ed85424a0e149fafb244f27\": container with ID starting with 2a81fec41e74683784c79d39df61a55233b276093ed85424a0e149fafb244f27 not found: ID does not exist" Feb 25 14:18:10 crc kubenswrapper[4815]: I0225 14:18:10.945465 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82c1c12e-5e32-440c-a6fd-c33486abd377" path="/var/lib/kubelet/pods/82c1c12e-5e32-440c-a6fd-c33486abd377/volumes" Feb 25 14:18:12 crc kubenswrapper[4815]: I0225 14:18:12.664265 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-xf9wz_bc773872-e8b3-4d04-acb3-593cb469b261/cert-manager-controller/0.log" Feb 25 14:18:12 crc kubenswrapper[4815]: I0225 14:18:12.741897 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-hxclh_ecdcd2c0-4b94-44aa-91fe-0fd1b09909f1/cert-manager-cainjector/0.log" Feb 25 14:18:12 crc kubenswrapper[4815]: I0225 14:18:12.861173 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-qslns_7b298115-612a-4691-8e63-067ce1230fdf/cert-manager-webhook/0.log" Feb 25 14:18:25 crc kubenswrapper[4815]: I0225 14:18:25.687089 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5c78fc5d65-wbxlg_bb89c9f3-b457-466d-a6a5-799ab144b125/nmstate-console-plugin/0.log" Feb 25 14:18:26 crc kubenswrapper[4815]: I0225 14:18:26.034025 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-p9cpt_43c7d725-8292-4e17-bc55-bfb18a7bc51e/nmstate-handler/0.log" Feb 25 14:18:26 crc kubenswrapper[4815]: I0225 14:18:26.050812 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-ljnd9_40f20bd3-e16c-490e-9316-fa6c6e36ab84/kube-rbac-proxy/0.log" Feb 25 14:18:26 crc kubenswrapper[4815]: I0225 14:18:26.166583 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-ljnd9_40f20bd3-e16c-490e-9316-fa6c6e36ab84/nmstate-metrics/0.log" Feb 25 14:18:26 crc kubenswrapper[4815]: I0225 14:18:26.257709 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-694c9596b7-xwh42_ffbc3f3b-118b-4b77-b3a1-b8b20201a08b/nmstate-operator/0.log" Feb 25 14:18:26 crc kubenswrapper[4815]: I0225 14:18:26.387614 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-866bcb46dc-sgk4q_266d6bc5-8e54-4af2-b6e4-5b3884ab882e/nmstate-webhook/0.log" Feb 25 14:18:49 crc kubenswrapper[4815]: I0225 14:18:49.931266 4815 scope.go:117] "RemoveContainer" containerID="9eff1e11700a748ecb9c048ad75b40ddead75959e9d30123375712977a42f6b3" Feb 25 14:18:53 crc kubenswrapper[4815]: I0225 14:18:53.164488 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-f6fqm_1b6c9556-1980-4581-8de1-f1fbf5d8256a/kube-rbac-proxy/0.log" Feb 25 14:18:53 crc kubenswrapper[4815]: I0225 14:18:53.240399 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-f6fqm_1b6c9556-1980-4581-8de1-f1fbf5d8256a/controller/0.log" Feb 25 14:18:53 crc kubenswrapper[4815]: I0225 14:18:53.388159 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/cp-frr-files/0.log" Feb 25 14:18:53 crc kubenswrapper[4815]: I0225 14:18:53.509465 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/cp-frr-files/0.log" Feb 25 14:18:53 crc kubenswrapper[4815]: I0225 14:18:53.521995 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/cp-reloader/0.log" Feb 25 14:18:53 crc kubenswrapper[4815]: I0225 14:18:53.525954 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/cp-metrics/0.log" Feb 25 14:18:53 crc kubenswrapper[4815]: I0225 14:18:53.588448 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/cp-reloader/0.log" Feb 25 14:18:53 crc kubenswrapper[4815]: I0225 14:18:53.713142 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/cp-frr-files/0.log" Feb 25 14:18:53 crc kubenswrapper[4815]: I0225 14:18:53.749119 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/cp-metrics/0.log" Feb 25 14:18:53 crc kubenswrapper[4815]: I0225 14:18:53.769680 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/cp-reloader/0.log" Feb 25 14:18:53 crc kubenswrapper[4815]: I0225 14:18:53.801262 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/cp-metrics/0.log" Feb 25 14:18:53 crc kubenswrapper[4815]: I0225 14:18:53.936671 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/cp-frr-files/0.log" Feb 25 14:18:53 crc kubenswrapper[4815]: I0225 14:18:53.973770 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/cp-reloader/0.log" Feb 25 14:18:53 crc kubenswrapper[4815]: I0225 14:18:53.993763 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/cp-metrics/0.log" Feb 25 14:18:54 crc kubenswrapper[4815]: I0225 14:18:54.023697 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/controller/0.log" Feb 25 14:18:54 crc kubenswrapper[4815]: I0225 14:18:54.165570 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/frr-metrics/0.log" Feb 25 14:18:54 crc kubenswrapper[4815]: I0225 14:18:54.235832 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/kube-rbac-proxy-frr/0.log" Feb 25 14:18:54 crc kubenswrapper[4815]: I0225 14:18:54.241678 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/kube-rbac-proxy/0.log" Feb 25 14:18:54 crc kubenswrapper[4815]: I0225 14:18:54.373616 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/reloader/0.log" Feb 25 14:18:54 crc kubenswrapper[4815]: I0225 14:18:54.447657 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-mcphd_699a6712-3fed-4b60-a2ea-e7440dd0e306/frr-k8s-webhook-server/0.log" Feb 25 14:18:54 crc kubenswrapper[4815]: I0225 14:18:54.728275 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-8674bc4d66-cm4rv_e0c4f375-94ec-47b4-885c-0b86b980580e/manager/0.log" Feb 25 14:18:54 crc kubenswrapper[4815]: I0225 14:18:54.877655 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5b77f59548-v7w5b_2f7b411f-fe90-48bf-995b-a40a6a8357ee/webhook-server/0.log" Feb 25 14:18:54 crc kubenswrapper[4815]: I0225 14:18:54.954819 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qxvv8_cefeaa51-3ff5-4124-8e3d-879d579f0d91/kube-rbac-proxy/0.log" Feb 25 14:18:55 crc kubenswrapper[4815]: I0225 14:18:55.570960 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qxvv8_cefeaa51-3ff5-4124-8e3d-879d579f0d91/speaker/0.log" Feb 25 14:18:55 crc kubenswrapper[4815]: I0225 14:18:55.830479 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-92bcq_801eb953-96a0-4272-9f81-13e80e61f75a/frr/0.log" Feb 25 14:19:07 crc kubenswrapper[4815]: I0225 14:19:07.957668 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx_8a6513b5-471c-4fb3-bdd7-4cf5e1528db5/util/0.log" Feb 25 14:19:08 crc kubenswrapper[4815]: I0225 14:19:08.144553 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx_8a6513b5-471c-4fb3-bdd7-4cf5e1528db5/util/0.log" Feb 25 14:19:08 crc kubenswrapper[4815]: I0225 14:19:08.153285 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx_8a6513b5-471c-4fb3-bdd7-4cf5e1528db5/pull/0.log" Feb 25 14:19:08 crc kubenswrapper[4815]: I0225 14:19:08.165810 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx_8a6513b5-471c-4fb3-bdd7-4cf5e1528db5/pull/0.log" Feb 25 14:19:08 crc kubenswrapper[4815]: I0225 14:19:08.316667 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx_8a6513b5-471c-4fb3-bdd7-4cf5e1528db5/extract/0.log" Feb 25 14:19:08 crc kubenswrapper[4815]: I0225 14:19:08.343394 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx_8a6513b5-471c-4fb3-bdd7-4cf5e1528db5/util/0.log" Feb 25 14:19:08 crc kubenswrapper[4815]: I0225 14:19:08.346716 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213p7tsx_8a6513b5-471c-4fb3-bdd7-4cf5e1528db5/pull/0.log" Feb 25 14:19:08 crc kubenswrapper[4815]: I0225 14:19:08.486997 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dsjpl_0781a2e4-ca50-4a4a-b35d-d4e448e99f9c/extract-utilities/0.log" Feb 25 14:19:08 crc kubenswrapper[4815]: I0225 14:19:08.675122 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dsjpl_0781a2e4-ca50-4a4a-b35d-d4e448e99f9c/extract-content/0.log" Feb 25 14:19:08 crc kubenswrapper[4815]: I0225 14:19:08.705378 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dsjpl_0781a2e4-ca50-4a4a-b35d-d4e448e99f9c/extract-content/0.log" Feb 25 14:19:08 crc kubenswrapper[4815]: I0225 14:19:08.710208 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dsjpl_0781a2e4-ca50-4a4a-b35d-d4e448e99f9c/extract-utilities/0.log" Feb 25 14:19:08 crc kubenswrapper[4815]: I0225 14:19:08.864015 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dsjpl_0781a2e4-ca50-4a4a-b35d-d4e448e99f9c/extract-utilities/0.log" Feb 25 14:19:08 crc kubenswrapper[4815]: I0225 14:19:08.965864 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dsjpl_0781a2e4-ca50-4a4a-b35d-d4e448e99f9c/extract-content/0.log" Feb 25 14:19:09 crc kubenswrapper[4815]: I0225 14:19:09.106391 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z6fgs_347a4d3a-d914-42b3-9239-ed7526039de6/extract-utilities/0.log" Feb 25 14:19:09 crc kubenswrapper[4815]: I0225 14:19:09.237789 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dsjpl_0781a2e4-ca50-4a4a-b35d-d4e448e99f9c/registry-server/0.log" Feb 25 14:19:09 crc kubenswrapper[4815]: I0225 14:19:09.289642 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z6fgs_347a4d3a-d914-42b3-9239-ed7526039de6/extract-utilities/0.log" Feb 25 14:19:09 crc kubenswrapper[4815]: I0225 14:19:09.290969 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z6fgs_347a4d3a-d914-42b3-9239-ed7526039de6/extract-content/0.log" Feb 25 14:19:09 crc kubenswrapper[4815]: I0225 14:19:09.341106 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z6fgs_347a4d3a-d914-42b3-9239-ed7526039de6/extract-content/0.log" Feb 25 14:19:09 crc kubenswrapper[4815]: I0225 14:19:09.554912 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z6fgs_347a4d3a-d914-42b3-9239-ed7526039de6/extract-content/0.log" Feb 25 14:19:09 crc kubenswrapper[4815]: I0225 14:19:09.568403 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z6fgs_347a4d3a-d914-42b3-9239-ed7526039de6/extract-utilities/0.log" Feb 25 14:19:09 crc kubenswrapper[4815]: I0225 14:19:09.764262 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp_6702f60b-0e49-4104-9434-6c3bf4e77cae/util/0.log" Feb 25 14:19:09 crc kubenswrapper[4815]: I0225 14:19:09.969938 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z6fgs_347a4d3a-d914-42b3-9239-ed7526039de6/registry-server/0.log" Feb 25 14:19:10 crc kubenswrapper[4815]: I0225 14:19:10.017353 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp_6702f60b-0e49-4104-9434-6c3bf4e77cae/pull/0.log" Feb 25 14:19:10 crc kubenswrapper[4815]: I0225 14:19:10.022162 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp_6702f60b-0e49-4104-9434-6c3bf4e77cae/util/0.log" Feb 25 14:19:10 crc kubenswrapper[4815]: I0225 14:19:10.068277 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp_6702f60b-0e49-4104-9434-6c3bf4e77cae/pull/0.log" Feb 25 14:19:10 crc kubenswrapper[4815]: I0225 14:19:10.170867 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp_6702f60b-0e49-4104-9434-6c3bf4e77cae/util/0.log" Feb 25 14:19:10 crc kubenswrapper[4815]: I0225 14:19:10.198674 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp_6702f60b-0e49-4104-9434-6c3bf4e77cae/pull/0.log" Feb 25 14:19:10 crc kubenswrapper[4815]: I0225 14:19:10.200705 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecarxcsp_6702f60b-0e49-4104-9434-6c3bf4e77cae/extract/0.log" Feb 25 14:19:10 crc kubenswrapper[4815]: I0225 14:19:10.379237 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8shd_06ec1b5e-fc51-4fe5-83d0-9697adaf1c57/extract-utilities/0.log" Feb 25 14:19:10 crc kubenswrapper[4815]: I0225 14:19:10.439140 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-ltn9g_4685a3d5-0377-4fb7-a496-854c23dd3729/marketplace-operator/0.log" Feb 25 14:19:10 crc kubenswrapper[4815]: I0225 14:19:10.623781 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8shd_06ec1b5e-fc51-4fe5-83d0-9697adaf1c57/extract-utilities/0.log" Feb 25 14:19:10 crc kubenswrapper[4815]: I0225 14:19:10.659959 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8shd_06ec1b5e-fc51-4fe5-83d0-9697adaf1c57/extract-content/0.log" Feb 25 14:19:10 crc kubenswrapper[4815]: I0225 14:19:10.672897 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8shd_06ec1b5e-fc51-4fe5-83d0-9697adaf1c57/extract-content/0.log" Feb 25 14:19:10 crc kubenswrapper[4815]: I0225 14:19:10.822379 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8shd_06ec1b5e-fc51-4fe5-83d0-9697adaf1c57/extract-content/0.log" Feb 25 14:19:10 crc kubenswrapper[4815]: I0225 14:19:10.861645 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8shd_06ec1b5e-fc51-4fe5-83d0-9697adaf1c57/extract-utilities/0.log" Feb 25 14:19:10 crc kubenswrapper[4815]: I0225 14:19:10.956774 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8shd_06ec1b5e-fc51-4fe5-83d0-9697adaf1c57/registry-server/0.log" Feb 25 14:19:11 crc kubenswrapper[4815]: I0225 14:19:11.032112 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xw2zf_9db66c63-07f2-4e45-851e-108477a7159d/extract-utilities/0.log" Feb 25 14:19:11 crc kubenswrapper[4815]: I0225 14:19:11.171267 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xw2zf_9db66c63-07f2-4e45-851e-108477a7159d/extract-utilities/0.log" Feb 25 14:19:11 crc kubenswrapper[4815]: I0225 14:19:11.200268 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xw2zf_9db66c63-07f2-4e45-851e-108477a7159d/extract-content/0.log" Feb 25 14:19:11 crc kubenswrapper[4815]: I0225 14:19:11.242026 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xw2zf_9db66c63-07f2-4e45-851e-108477a7159d/extract-content/0.log" Feb 25 14:19:11 crc kubenswrapper[4815]: I0225 14:19:11.408144 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xw2zf_9db66c63-07f2-4e45-851e-108477a7159d/extract-content/0.log" Feb 25 14:19:11 crc kubenswrapper[4815]: I0225 14:19:11.527108 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xw2zf_9db66c63-07f2-4e45-851e-108477a7159d/extract-utilities/0.log" Feb 25 14:19:12 crc kubenswrapper[4815]: I0225 14:19:12.009597 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xw2zf_9db66c63-07f2-4e45-851e-108477a7159d/registry-server/0.log" Feb 25 14:19:51 crc kubenswrapper[4815]: I0225 14:19:51.342353 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 14:19:51 crc kubenswrapper[4815]: I0225 14:19:51.343248 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 14:20:00 crc kubenswrapper[4815]: I0225 14:20:00.150784 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533820-7bkpp"] Feb 25 14:20:00 crc kubenswrapper[4815]: E0225 14:20:00.151869 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c1c12e-5e32-440c-a6fd-c33486abd377" containerName="registry-server" Feb 25 14:20:00 crc kubenswrapper[4815]: I0225 14:20:00.151888 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c1c12e-5e32-440c-a6fd-c33486abd377" containerName="registry-server" Feb 25 14:20:00 crc kubenswrapper[4815]: E0225 14:20:00.151911 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23253ded-9aae-451f-a7fc-1b10518f4312" containerName="oc" Feb 25 14:20:00 crc kubenswrapper[4815]: I0225 14:20:00.151922 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="23253ded-9aae-451f-a7fc-1b10518f4312" containerName="oc" Feb 25 14:20:00 crc kubenswrapper[4815]: E0225 14:20:00.151945 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c1c12e-5e32-440c-a6fd-c33486abd377" containerName="extract-utilities" Feb 25 14:20:00 crc kubenswrapper[4815]: I0225 14:20:00.151957 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c1c12e-5e32-440c-a6fd-c33486abd377" containerName="extract-utilities" Feb 25 14:20:00 crc kubenswrapper[4815]: E0225 14:20:00.151995 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c1c12e-5e32-440c-a6fd-c33486abd377" containerName="extract-content" Feb 25 14:20:00 crc kubenswrapper[4815]: I0225 14:20:00.152004 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c1c12e-5e32-440c-a6fd-c33486abd377" containerName="extract-content" Feb 25 14:20:00 crc kubenswrapper[4815]: I0225 14:20:00.152228 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="82c1c12e-5e32-440c-a6fd-c33486abd377" containerName="registry-server" Feb 25 14:20:00 crc kubenswrapper[4815]: I0225 14:20:00.152247 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="23253ded-9aae-451f-a7fc-1b10518f4312" containerName="oc" Feb 25 14:20:00 crc kubenswrapper[4815]: I0225 14:20:00.153101 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533820-7bkpp" Feb 25 14:20:00 crc kubenswrapper[4815]: I0225 14:20:00.159533 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533820-7bkpp"] Feb 25 14:20:00 crc kubenswrapper[4815]: I0225 14:20:00.161249 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 14:20:00 crc kubenswrapper[4815]: I0225 14:20:00.164347 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 14:20:00 crc kubenswrapper[4815]: I0225 14:20:00.164428 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 14:20:00 crc kubenswrapper[4815]: I0225 14:20:00.259908 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-649vr\" (UniqueName: \"kubernetes.io/projected/fbddfcbe-0ae8-4d91-a8d8-7d8314849143-kube-api-access-649vr\") pod \"auto-csr-approver-29533820-7bkpp\" (UID: \"fbddfcbe-0ae8-4d91-a8d8-7d8314849143\") " pod="openshift-infra/auto-csr-approver-29533820-7bkpp" Feb 25 14:20:00 crc kubenswrapper[4815]: I0225 14:20:00.362433 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-649vr\" (UniqueName: \"kubernetes.io/projected/fbddfcbe-0ae8-4d91-a8d8-7d8314849143-kube-api-access-649vr\") pod \"auto-csr-approver-29533820-7bkpp\" (UID: \"fbddfcbe-0ae8-4d91-a8d8-7d8314849143\") " pod="openshift-infra/auto-csr-approver-29533820-7bkpp" Feb 25 14:20:00 crc kubenswrapper[4815]: I0225 14:20:00.391668 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-649vr\" (UniqueName: \"kubernetes.io/projected/fbddfcbe-0ae8-4d91-a8d8-7d8314849143-kube-api-access-649vr\") pod \"auto-csr-approver-29533820-7bkpp\" (UID: \"fbddfcbe-0ae8-4d91-a8d8-7d8314849143\") " pod="openshift-infra/auto-csr-approver-29533820-7bkpp" Feb 25 14:20:00 crc kubenswrapper[4815]: I0225 14:20:00.484008 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533820-7bkpp" Feb 25 14:20:00 crc kubenswrapper[4815]: I0225 14:20:00.949413 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533820-7bkpp"] Feb 25 14:20:00 crc kubenswrapper[4815]: W0225 14:20:00.957178 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfbddfcbe_0ae8_4d91_a8d8_7d8314849143.slice/crio-fa8b42cf851741143b7d60c59d85ae8cc09c793aba14e61065fd0d6bd7b11e0d WatchSource:0}: Error finding container fa8b42cf851741143b7d60c59d85ae8cc09c793aba14e61065fd0d6bd7b11e0d: Status 404 returned error can't find the container with id fa8b42cf851741143b7d60c59d85ae8cc09c793aba14e61065fd0d6bd7b11e0d Feb 25 14:20:01 crc kubenswrapper[4815]: I0225 14:20:01.059240 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533820-7bkpp" event={"ID":"fbddfcbe-0ae8-4d91-a8d8-7d8314849143","Type":"ContainerStarted","Data":"fa8b42cf851741143b7d60c59d85ae8cc09c793aba14e61065fd0d6bd7b11e0d"} Feb 25 14:20:03 crc kubenswrapper[4815]: I0225 14:20:03.100700 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533820-7bkpp" podStartSLOduration=1.250143552 podStartE2EDuration="3.100685729s" podCreationTimestamp="2026-02-25 14:20:00 +0000 UTC" firstStartedPulling="2026-02-25 14:20:00.959692525 +0000 UTC m=+3558.760790579" lastFinishedPulling="2026-02-25 14:20:02.810234702 +0000 UTC m=+3560.611332756" observedRunningTime="2026-02-25 14:20:03.097618563 +0000 UTC m=+3560.898716607" watchObservedRunningTime="2026-02-25 14:20:03.100685729 +0000 UTC m=+3560.901783783" Feb 25 14:20:04 crc kubenswrapper[4815]: I0225 14:20:04.098577 4815 generic.go:334] "Generic (PLEG): container finished" podID="fbddfcbe-0ae8-4d91-a8d8-7d8314849143" containerID="a00b2f27fcab78b9a064f917480a60e47bdd039de2649e0f7721bbbe4ff69466" exitCode=0 Feb 25 14:20:04 crc kubenswrapper[4815]: I0225 14:20:04.099030 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533820-7bkpp" event={"ID":"fbddfcbe-0ae8-4d91-a8d8-7d8314849143","Type":"ContainerDied","Data":"a00b2f27fcab78b9a064f917480a60e47bdd039de2649e0f7721bbbe4ff69466"} Feb 25 14:20:05 crc kubenswrapper[4815]: I0225 14:20:05.531610 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533820-7bkpp" Feb 25 14:20:05 crc kubenswrapper[4815]: I0225 14:20:05.668222 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-649vr\" (UniqueName: \"kubernetes.io/projected/fbddfcbe-0ae8-4d91-a8d8-7d8314849143-kube-api-access-649vr\") pod \"fbddfcbe-0ae8-4d91-a8d8-7d8314849143\" (UID: \"fbddfcbe-0ae8-4d91-a8d8-7d8314849143\") " Feb 25 14:20:05 crc kubenswrapper[4815]: I0225 14:20:05.674122 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbddfcbe-0ae8-4d91-a8d8-7d8314849143-kube-api-access-649vr" (OuterVolumeSpecName: "kube-api-access-649vr") pod "fbddfcbe-0ae8-4d91-a8d8-7d8314849143" (UID: "fbddfcbe-0ae8-4d91-a8d8-7d8314849143"). InnerVolumeSpecName "kube-api-access-649vr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:20:05 crc kubenswrapper[4815]: I0225 14:20:05.770572 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-649vr\" (UniqueName: \"kubernetes.io/projected/fbddfcbe-0ae8-4d91-a8d8-7d8314849143-kube-api-access-649vr\") on node \"crc\" DevicePath \"\"" Feb 25 14:20:06 crc kubenswrapper[4815]: I0225 14:20:06.031127 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533814-mjcm4"] Feb 25 14:20:06 crc kubenswrapper[4815]: I0225 14:20:06.039615 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533814-mjcm4"] Feb 25 14:20:06 crc kubenswrapper[4815]: I0225 14:20:06.124434 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533820-7bkpp" event={"ID":"fbddfcbe-0ae8-4d91-a8d8-7d8314849143","Type":"ContainerDied","Data":"fa8b42cf851741143b7d60c59d85ae8cc09c793aba14e61065fd0d6bd7b11e0d"} Feb 25 14:20:06 crc kubenswrapper[4815]: I0225 14:20:06.124482 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa8b42cf851741143b7d60c59d85ae8cc09c793aba14e61065fd0d6bd7b11e0d" Feb 25 14:20:06 crc kubenswrapper[4815]: I0225 14:20:06.124534 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533820-7bkpp" Feb 25 14:20:06 crc kubenswrapper[4815]: I0225 14:20:06.947872 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1e09176-fa9b-4ace-bb60-ee5f54de9571" path="/var/lib/kubelet/pods/f1e09176-fa9b-4ace-bb60-ee5f54de9571/volumes" Feb 25 14:20:21 crc kubenswrapper[4815]: I0225 14:20:21.342249 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 14:20:21 crc kubenswrapper[4815]: I0225 14:20:21.342843 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 14:20:50 crc kubenswrapper[4815]: I0225 14:20:50.077497 4815 scope.go:117] "RemoveContainer" containerID="a4b9a30738cf19ad63e8a58ea4fd11a35641d4d44704d0f3755d1ca1d18aed46" Feb 25 14:20:51 crc kubenswrapper[4815]: I0225 14:20:51.342149 4815 patch_prober.go:28] interesting pod/machine-config-daemon-frc75 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 14:20:51 crc kubenswrapper[4815]: I0225 14:20:51.342583 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 14:20:51 crc kubenswrapper[4815]: I0225 14:20:51.342656 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frc75" Feb 25 14:20:51 crc kubenswrapper[4815]: I0225 14:20:51.343572 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107"} pod="openshift-machine-config-operator/machine-config-daemon-frc75" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 14:20:51 crc kubenswrapper[4815]: I0225 14:20:51.343625 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerName="machine-config-daemon" containerID="cri-o://8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" gracePeriod=600 Feb 25 14:20:51 crc kubenswrapper[4815]: E0225 14:20:51.477981 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:20:51 crc kubenswrapper[4815]: I0225 14:20:51.579687 4815 generic.go:334] "Generic (PLEG): container finished" podID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" exitCode=0 Feb 25 14:20:51 crc kubenswrapper[4815]: I0225 14:20:51.579743 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerDied","Data":"8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107"} Feb 25 14:20:51 crc kubenswrapper[4815]: I0225 14:20:51.579784 4815 scope.go:117] "RemoveContainer" containerID="35ad0ed6ac05ca7660eb380c1f3eb06169b51f834d4f9b66e462803955bc5ff9" Feb 25 14:20:51 crc kubenswrapper[4815]: I0225 14:20:51.580893 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:20:51 crc kubenswrapper[4815]: E0225 14:20:51.581277 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:20:58 crc kubenswrapper[4815]: I0225 14:20:58.656839 4815 generic.go:334] "Generic (PLEG): container finished" podID="8da95553-a7bd-4cd4-84ea-10a8acf68ced" containerID="6e262e4a7a08ff1daa2e69e8c3427adfc036dfff10ae58f297932ffa13e89dd9" exitCode=0 Feb 25 14:20:58 crc kubenswrapper[4815]: I0225 14:20:58.656934 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-twd5c/must-gather-v7wxh" event={"ID":"8da95553-a7bd-4cd4-84ea-10a8acf68ced","Type":"ContainerDied","Data":"6e262e4a7a08ff1daa2e69e8c3427adfc036dfff10ae58f297932ffa13e89dd9"} Feb 25 14:20:58 crc kubenswrapper[4815]: I0225 14:20:58.657897 4815 scope.go:117] "RemoveContainer" containerID="6e262e4a7a08ff1daa2e69e8c3427adfc036dfff10ae58f297932ffa13e89dd9" Feb 25 14:20:58 crc kubenswrapper[4815]: I0225 14:20:58.933301 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-twd5c_must-gather-v7wxh_8da95553-a7bd-4cd4-84ea-10a8acf68ced/gather/0.log" Feb 25 14:21:04 crc kubenswrapper[4815]: I0225 14:21:04.936224 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:21:04 crc kubenswrapper[4815]: E0225 14:21:04.937309 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:21:06 crc kubenswrapper[4815]: I0225 14:21:06.853100 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-twd5c/must-gather-v7wxh"] Feb 25 14:21:06 crc kubenswrapper[4815]: I0225 14:21:06.853731 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-twd5c/must-gather-v7wxh" podUID="8da95553-a7bd-4cd4-84ea-10a8acf68ced" containerName="copy" containerID="cri-o://11f64d995172c2029009adb72beb11cd88f94fd2d23a9d47d1839d0c5bc25625" gracePeriod=2 Feb 25 14:21:06 crc kubenswrapper[4815]: I0225 14:21:06.865916 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-twd5c/must-gather-v7wxh"] Feb 25 14:21:07 crc kubenswrapper[4815]: I0225 14:21:07.294675 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-twd5c_must-gather-v7wxh_8da95553-a7bd-4cd4-84ea-10a8acf68ced/copy/0.log" Feb 25 14:21:07 crc kubenswrapper[4815]: I0225 14:21:07.295357 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-twd5c/must-gather-v7wxh" Feb 25 14:21:07 crc kubenswrapper[4815]: I0225 14:21:07.335365 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8da95553-a7bd-4cd4-84ea-10a8acf68ced-must-gather-output\") pod \"8da95553-a7bd-4cd4-84ea-10a8acf68ced\" (UID: \"8da95553-a7bd-4cd4-84ea-10a8acf68ced\") " Feb 25 14:21:07 crc kubenswrapper[4815]: I0225 14:21:07.335488 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szbq8\" (UniqueName: \"kubernetes.io/projected/8da95553-a7bd-4cd4-84ea-10a8acf68ced-kube-api-access-szbq8\") pod \"8da95553-a7bd-4cd4-84ea-10a8acf68ced\" (UID: \"8da95553-a7bd-4cd4-84ea-10a8acf68ced\") " Feb 25 14:21:07 crc kubenswrapper[4815]: I0225 14:21:07.343654 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8da95553-a7bd-4cd4-84ea-10a8acf68ced-kube-api-access-szbq8" (OuterVolumeSpecName: "kube-api-access-szbq8") pod "8da95553-a7bd-4cd4-84ea-10a8acf68ced" (UID: "8da95553-a7bd-4cd4-84ea-10a8acf68ced"). InnerVolumeSpecName "kube-api-access-szbq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:21:07 crc kubenswrapper[4815]: I0225 14:21:07.437279 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szbq8\" (UniqueName: \"kubernetes.io/projected/8da95553-a7bd-4cd4-84ea-10a8acf68ced-kube-api-access-szbq8\") on node \"crc\" DevicePath \"\"" Feb 25 14:21:07 crc kubenswrapper[4815]: I0225 14:21:07.491102 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8da95553-a7bd-4cd4-84ea-10a8acf68ced-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "8da95553-a7bd-4cd4-84ea-10a8acf68ced" (UID: "8da95553-a7bd-4cd4-84ea-10a8acf68ced"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:21:07 crc kubenswrapper[4815]: I0225 14:21:07.539207 4815 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8da95553-a7bd-4cd4-84ea-10a8acf68ced-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 25 14:21:07 crc kubenswrapper[4815]: I0225 14:21:07.744585 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-twd5c_must-gather-v7wxh_8da95553-a7bd-4cd4-84ea-10a8acf68ced/copy/0.log" Feb 25 14:21:07 crc kubenswrapper[4815]: I0225 14:21:07.745313 4815 generic.go:334] "Generic (PLEG): container finished" podID="8da95553-a7bd-4cd4-84ea-10a8acf68ced" containerID="11f64d995172c2029009adb72beb11cd88f94fd2d23a9d47d1839d0c5bc25625" exitCode=143 Feb 25 14:21:07 crc kubenswrapper[4815]: I0225 14:21:07.745371 4815 scope.go:117] "RemoveContainer" containerID="11f64d995172c2029009adb72beb11cd88f94fd2d23a9d47d1839d0c5bc25625" Feb 25 14:21:07 crc kubenswrapper[4815]: I0225 14:21:07.745374 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-twd5c/must-gather-v7wxh" Feb 25 14:21:07 crc kubenswrapper[4815]: I0225 14:21:07.768223 4815 scope.go:117] "RemoveContainer" containerID="6e262e4a7a08ff1daa2e69e8c3427adfc036dfff10ae58f297932ffa13e89dd9" Feb 25 14:21:07 crc kubenswrapper[4815]: I0225 14:21:07.841939 4815 scope.go:117] "RemoveContainer" containerID="11f64d995172c2029009adb72beb11cd88f94fd2d23a9d47d1839d0c5bc25625" Feb 25 14:21:07 crc kubenswrapper[4815]: E0225 14:21:07.842332 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11f64d995172c2029009adb72beb11cd88f94fd2d23a9d47d1839d0c5bc25625\": container with ID starting with 11f64d995172c2029009adb72beb11cd88f94fd2d23a9d47d1839d0c5bc25625 not found: ID does not exist" containerID="11f64d995172c2029009adb72beb11cd88f94fd2d23a9d47d1839d0c5bc25625" Feb 25 14:21:07 crc kubenswrapper[4815]: I0225 14:21:07.842392 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11f64d995172c2029009adb72beb11cd88f94fd2d23a9d47d1839d0c5bc25625"} err="failed to get container status \"11f64d995172c2029009adb72beb11cd88f94fd2d23a9d47d1839d0c5bc25625\": rpc error: code = NotFound desc = could not find container \"11f64d995172c2029009adb72beb11cd88f94fd2d23a9d47d1839d0c5bc25625\": container with ID starting with 11f64d995172c2029009adb72beb11cd88f94fd2d23a9d47d1839d0c5bc25625 not found: ID does not exist" Feb 25 14:21:07 crc kubenswrapper[4815]: I0225 14:21:07.842420 4815 scope.go:117] "RemoveContainer" containerID="6e262e4a7a08ff1daa2e69e8c3427adfc036dfff10ae58f297932ffa13e89dd9" Feb 25 14:21:07 crc kubenswrapper[4815]: E0225 14:21:07.842938 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e262e4a7a08ff1daa2e69e8c3427adfc036dfff10ae58f297932ffa13e89dd9\": container with ID starting with 6e262e4a7a08ff1daa2e69e8c3427adfc036dfff10ae58f297932ffa13e89dd9 not found: ID does not exist" containerID="6e262e4a7a08ff1daa2e69e8c3427adfc036dfff10ae58f297932ffa13e89dd9" Feb 25 14:21:07 crc kubenswrapper[4815]: I0225 14:21:07.842966 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e262e4a7a08ff1daa2e69e8c3427adfc036dfff10ae58f297932ffa13e89dd9"} err="failed to get container status \"6e262e4a7a08ff1daa2e69e8c3427adfc036dfff10ae58f297932ffa13e89dd9\": rpc error: code = NotFound desc = could not find container \"6e262e4a7a08ff1daa2e69e8c3427adfc036dfff10ae58f297932ffa13e89dd9\": container with ID starting with 6e262e4a7a08ff1daa2e69e8c3427adfc036dfff10ae58f297932ffa13e89dd9 not found: ID does not exist" Feb 25 14:21:08 crc kubenswrapper[4815]: I0225 14:21:08.945677 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8da95553-a7bd-4cd4-84ea-10a8acf68ced" path="/var/lib/kubelet/pods/8da95553-a7bd-4cd4-84ea-10a8acf68ced/volumes" Feb 25 14:21:16 crc kubenswrapper[4815]: I0225 14:21:16.937985 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:21:16 crc kubenswrapper[4815]: E0225 14:21:16.938841 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:21:28 crc kubenswrapper[4815]: I0225 14:21:28.935799 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:21:28 crc kubenswrapper[4815]: E0225 14:21:28.936779 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:21:40 crc kubenswrapper[4815]: I0225 14:21:40.937335 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:21:40 crc kubenswrapper[4815]: E0225 14:21:40.941379 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:21:42 crc kubenswrapper[4815]: I0225 14:21:42.816230 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4zmql"] Feb 25 14:21:42 crc kubenswrapper[4815]: E0225 14:21:42.817274 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbddfcbe-0ae8-4d91-a8d8-7d8314849143" containerName="oc" Feb 25 14:21:42 crc kubenswrapper[4815]: I0225 14:21:42.817298 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbddfcbe-0ae8-4d91-a8d8-7d8314849143" containerName="oc" Feb 25 14:21:42 crc kubenswrapper[4815]: E0225 14:21:42.817349 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8da95553-a7bd-4cd4-84ea-10a8acf68ced" containerName="gather" Feb 25 14:21:42 crc kubenswrapper[4815]: I0225 14:21:42.817360 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8da95553-a7bd-4cd4-84ea-10a8acf68ced" containerName="gather" Feb 25 14:21:42 crc kubenswrapper[4815]: E0225 14:21:42.817388 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8da95553-a7bd-4cd4-84ea-10a8acf68ced" containerName="copy" Feb 25 14:21:42 crc kubenswrapper[4815]: I0225 14:21:42.817399 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8da95553-a7bd-4cd4-84ea-10a8acf68ced" containerName="copy" Feb 25 14:21:42 crc kubenswrapper[4815]: I0225 14:21:42.817728 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8da95553-a7bd-4cd4-84ea-10a8acf68ced" containerName="copy" Feb 25 14:21:42 crc kubenswrapper[4815]: I0225 14:21:42.817767 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbddfcbe-0ae8-4d91-a8d8-7d8314849143" containerName="oc" Feb 25 14:21:42 crc kubenswrapper[4815]: I0225 14:21:42.817795 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8da95553-a7bd-4cd4-84ea-10a8acf68ced" containerName="gather" Feb 25 14:21:42 crc kubenswrapper[4815]: I0225 14:21:42.820009 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4zmql" Feb 25 14:21:42 crc kubenswrapper[4815]: I0225 14:21:42.831644 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4zmql"] Feb 25 14:21:42 crc kubenswrapper[4815]: I0225 14:21:42.965946 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15dc1c43-85d2-44ee-92a5-7094b563550e-catalog-content\") pod \"certified-operators-4zmql\" (UID: \"15dc1c43-85d2-44ee-92a5-7094b563550e\") " pod="openshift-marketplace/certified-operators-4zmql" Feb 25 14:21:42 crc kubenswrapper[4815]: I0225 14:21:42.966022 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15dc1c43-85d2-44ee-92a5-7094b563550e-utilities\") pod \"certified-operators-4zmql\" (UID: \"15dc1c43-85d2-44ee-92a5-7094b563550e\") " pod="openshift-marketplace/certified-operators-4zmql" Feb 25 14:21:42 crc kubenswrapper[4815]: I0225 14:21:42.966116 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6jmv\" (UniqueName: \"kubernetes.io/projected/15dc1c43-85d2-44ee-92a5-7094b563550e-kube-api-access-f6jmv\") pod \"certified-operators-4zmql\" (UID: \"15dc1c43-85d2-44ee-92a5-7094b563550e\") " pod="openshift-marketplace/certified-operators-4zmql" Feb 25 14:21:43 crc kubenswrapper[4815]: I0225 14:21:43.068887 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15dc1c43-85d2-44ee-92a5-7094b563550e-catalog-content\") pod \"certified-operators-4zmql\" (UID: \"15dc1c43-85d2-44ee-92a5-7094b563550e\") " pod="openshift-marketplace/certified-operators-4zmql" Feb 25 14:21:43 crc kubenswrapper[4815]: I0225 14:21:43.068956 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15dc1c43-85d2-44ee-92a5-7094b563550e-utilities\") pod \"certified-operators-4zmql\" (UID: \"15dc1c43-85d2-44ee-92a5-7094b563550e\") " pod="openshift-marketplace/certified-operators-4zmql" Feb 25 14:21:43 crc kubenswrapper[4815]: I0225 14:21:43.069002 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6jmv\" (UniqueName: \"kubernetes.io/projected/15dc1c43-85d2-44ee-92a5-7094b563550e-kube-api-access-f6jmv\") pod \"certified-operators-4zmql\" (UID: \"15dc1c43-85d2-44ee-92a5-7094b563550e\") " pod="openshift-marketplace/certified-operators-4zmql" Feb 25 14:21:43 crc kubenswrapper[4815]: I0225 14:21:43.069775 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15dc1c43-85d2-44ee-92a5-7094b563550e-catalog-content\") pod \"certified-operators-4zmql\" (UID: \"15dc1c43-85d2-44ee-92a5-7094b563550e\") " pod="openshift-marketplace/certified-operators-4zmql" Feb 25 14:21:43 crc kubenswrapper[4815]: I0225 14:21:43.069810 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15dc1c43-85d2-44ee-92a5-7094b563550e-utilities\") pod \"certified-operators-4zmql\" (UID: \"15dc1c43-85d2-44ee-92a5-7094b563550e\") " pod="openshift-marketplace/certified-operators-4zmql" Feb 25 14:21:43 crc kubenswrapper[4815]: I0225 14:21:43.088954 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6jmv\" (UniqueName: \"kubernetes.io/projected/15dc1c43-85d2-44ee-92a5-7094b563550e-kube-api-access-f6jmv\") pod \"certified-operators-4zmql\" (UID: \"15dc1c43-85d2-44ee-92a5-7094b563550e\") " pod="openshift-marketplace/certified-operators-4zmql" Feb 25 14:21:43 crc kubenswrapper[4815]: I0225 14:21:43.149026 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4zmql" Feb 25 14:21:43 crc kubenswrapper[4815]: I0225 14:21:43.667546 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4zmql"] Feb 25 14:21:44 crc kubenswrapper[4815]: I0225 14:21:44.120169 4815 generic.go:334] "Generic (PLEG): container finished" podID="15dc1c43-85d2-44ee-92a5-7094b563550e" containerID="151bb90b4e14e3597759cc3ae141ba73af0fc25992b21d8beb3f12e019735140" exitCode=0 Feb 25 14:21:44 crc kubenswrapper[4815]: I0225 14:21:44.120776 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4zmql" event={"ID":"15dc1c43-85d2-44ee-92a5-7094b563550e","Type":"ContainerDied","Data":"151bb90b4e14e3597759cc3ae141ba73af0fc25992b21d8beb3f12e019735140"} Feb 25 14:21:44 crc kubenswrapper[4815]: I0225 14:21:44.120858 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4zmql" event={"ID":"15dc1c43-85d2-44ee-92a5-7094b563550e","Type":"ContainerStarted","Data":"c7ad1e5491d172353c2e86139e92679441b92828d8dbd3b62e7c86754f1dfef7"} Feb 25 14:21:45 crc kubenswrapper[4815]: I0225 14:21:45.134265 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4zmql" event={"ID":"15dc1c43-85d2-44ee-92a5-7094b563550e","Type":"ContainerStarted","Data":"6c1bd96c0be4678656b17b82ae9cedb4da1d997e951d26c6355bbed515bf8d3d"} Feb 25 14:21:47 crc kubenswrapper[4815]: I0225 14:21:47.154032 4815 generic.go:334] "Generic (PLEG): container finished" podID="15dc1c43-85d2-44ee-92a5-7094b563550e" containerID="6c1bd96c0be4678656b17b82ae9cedb4da1d997e951d26c6355bbed515bf8d3d" exitCode=0 Feb 25 14:21:47 crc kubenswrapper[4815]: I0225 14:21:47.154094 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4zmql" event={"ID":"15dc1c43-85d2-44ee-92a5-7094b563550e","Type":"ContainerDied","Data":"6c1bd96c0be4678656b17b82ae9cedb4da1d997e951d26c6355bbed515bf8d3d"} Feb 25 14:21:48 crc kubenswrapper[4815]: I0225 14:21:48.168552 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4zmql" event={"ID":"15dc1c43-85d2-44ee-92a5-7094b563550e","Type":"ContainerStarted","Data":"98ad9145281ddac09a41d46dc7918222f69142044f0025e388edc7ad7d11cfb3"} Feb 25 14:21:48 crc kubenswrapper[4815]: I0225 14:21:48.195940 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4zmql" podStartSLOduration=2.694652177 podStartE2EDuration="6.195923483s" podCreationTimestamp="2026-02-25 14:21:42 +0000 UTC" firstStartedPulling="2026-02-25 14:21:44.122426877 +0000 UTC m=+3661.923524941" lastFinishedPulling="2026-02-25 14:21:47.623698193 +0000 UTC m=+3665.424796247" observedRunningTime="2026-02-25 14:21:48.193424915 +0000 UTC m=+3665.994522969" watchObservedRunningTime="2026-02-25 14:21:48.195923483 +0000 UTC m=+3665.997021537" Feb 25 14:21:53 crc kubenswrapper[4815]: I0225 14:21:53.149603 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4zmql" Feb 25 14:21:53 crc kubenswrapper[4815]: I0225 14:21:53.151076 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4zmql" Feb 25 14:21:53 crc kubenswrapper[4815]: I0225 14:21:53.204817 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4zmql" Feb 25 14:21:53 crc kubenswrapper[4815]: I0225 14:21:53.287500 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4zmql" Feb 25 14:21:53 crc kubenswrapper[4815]: I0225 14:21:53.471770 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4zmql"] Feb 25 14:21:54 crc kubenswrapper[4815]: I0225 14:21:54.935792 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:21:54 crc kubenswrapper[4815]: E0225 14:21:54.936640 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:21:55 crc kubenswrapper[4815]: I0225 14:21:55.251799 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4zmql" podUID="15dc1c43-85d2-44ee-92a5-7094b563550e" containerName="registry-server" containerID="cri-o://98ad9145281ddac09a41d46dc7918222f69142044f0025e388edc7ad7d11cfb3" gracePeriod=2 Feb 25 14:21:55 crc kubenswrapper[4815]: I0225 14:21:55.700720 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4zmql" Feb 25 14:21:55 crc kubenswrapper[4815]: I0225 14:21:55.851866 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6jmv\" (UniqueName: \"kubernetes.io/projected/15dc1c43-85d2-44ee-92a5-7094b563550e-kube-api-access-f6jmv\") pod \"15dc1c43-85d2-44ee-92a5-7094b563550e\" (UID: \"15dc1c43-85d2-44ee-92a5-7094b563550e\") " Feb 25 14:21:55 crc kubenswrapper[4815]: I0225 14:21:55.852044 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15dc1c43-85d2-44ee-92a5-7094b563550e-utilities\") pod \"15dc1c43-85d2-44ee-92a5-7094b563550e\" (UID: \"15dc1c43-85d2-44ee-92a5-7094b563550e\") " Feb 25 14:21:55 crc kubenswrapper[4815]: I0225 14:21:55.852074 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15dc1c43-85d2-44ee-92a5-7094b563550e-catalog-content\") pod \"15dc1c43-85d2-44ee-92a5-7094b563550e\" (UID: \"15dc1c43-85d2-44ee-92a5-7094b563550e\") " Feb 25 14:21:55 crc kubenswrapper[4815]: I0225 14:21:55.853993 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15dc1c43-85d2-44ee-92a5-7094b563550e-utilities" (OuterVolumeSpecName: "utilities") pod "15dc1c43-85d2-44ee-92a5-7094b563550e" (UID: "15dc1c43-85d2-44ee-92a5-7094b563550e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:21:55 crc kubenswrapper[4815]: I0225 14:21:55.866298 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15dc1c43-85d2-44ee-92a5-7094b563550e-kube-api-access-f6jmv" (OuterVolumeSpecName: "kube-api-access-f6jmv") pod "15dc1c43-85d2-44ee-92a5-7094b563550e" (UID: "15dc1c43-85d2-44ee-92a5-7094b563550e"). InnerVolumeSpecName "kube-api-access-f6jmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:21:55 crc kubenswrapper[4815]: I0225 14:21:55.934013 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15dc1c43-85d2-44ee-92a5-7094b563550e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15dc1c43-85d2-44ee-92a5-7094b563550e" (UID: "15dc1c43-85d2-44ee-92a5-7094b563550e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:21:55 crc kubenswrapper[4815]: I0225 14:21:55.954334 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6jmv\" (UniqueName: \"kubernetes.io/projected/15dc1c43-85d2-44ee-92a5-7094b563550e-kube-api-access-f6jmv\") on node \"crc\" DevicePath \"\"" Feb 25 14:21:55 crc kubenswrapper[4815]: I0225 14:21:55.954360 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15dc1c43-85d2-44ee-92a5-7094b563550e-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 14:21:55 crc kubenswrapper[4815]: I0225 14:21:55.954372 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15dc1c43-85d2-44ee-92a5-7094b563550e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 14:21:56 crc kubenswrapper[4815]: I0225 14:21:56.261406 4815 generic.go:334] "Generic (PLEG): container finished" podID="15dc1c43-85d2-44ee-92a5-7094b563550e" containerID="98ad9145281ddac09a41d46dc7918222f69142044f0025e388edc7ad7d11cfb3" exitCode=0 Feb 25 14:21:56 crc kubenswrapper[4815]: I0225 14:21:56.261447 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4zmql" event={"ID":"15dc1c43-85d2-44ee-92a5-7094b563550e","Type":"ContainerDied","Data":"98ad9145281ddac09a41d46dc7918222f69142044f0025e388edc7ad7d11cfb3"} Feb 25 14:21:56 crc kubenswrapper[4815]: I0225 14:21:56.261487 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4zmql" Feb 25 14:21:56 crc kubenswrapper[4815]: I0225 14:21:56.261521 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4zmql" event={"ID":"15dc1c43-85d2-44ee-92a5-7094b563550e","Type":"ContainerDied","Data":"c7ad1e5491d172353c2e86139e92679441b92828d8dbd3b62e7c86754f1dfef7"} Feb 25 14:21:56 crc kubenswrapper[4815]: I0225 14:21:56.261547 4815 scope.go:117] "RemoveContainer" containerID="98ad9145281ddac09a41d46dc7918222f69142044f0025e388edc7ad7d11cfb3" Feb 25 14:21:56 crc kubenswrapper[4815]: I0225 14:21:56.299085 4815 scope.go:117] "RemoveContainer" containerID="6c1bd96c0be4678656b17b82ae9cedb4da1d997e951d26c6355bbed515bf8d3d" Feb 25 14:21:56 crc kubenswrapper[4815]: I0225 14:21:56.301217 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4zmql"] Feb 25 14:21:56 crc kubenswrapper[4815]: I0225 14:21:56.311424 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4zmql"] Feb 25 14:21:56 crc kubenswrapper[4815]: I0225 14:21:56.323848 4815 scope.go:117] "RemoveContainer" containerID="151bb90b4e14e3597759cc3ae141ba73af0fc25992b21d8beb3f12e019735140" Feb 25 14:21:56 crc kubenswrapper[4815]: I0225 14:21:56.401757 4815 scope.go:117] "RemoveContainer" containerID="98ad9145281ddac09a41d46dc7918222f69142044f0025e388edc7ad7d11cfb3" Feb 25 14:21:56 crc kubenswrapper[4815]: E0225 14:21:56.402118 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98ad9145281ddac09a41d46dc7918222f69142044f0025e388edc7ad7d11cfb3\": container with ID starting with 98ad9145281ddac09a41d46dc7918222f69142044f0025e388edc7ad7d11cfb3 not found: ID does not exist" containerID="98ad9145281ddac09a41d46dc7918222f69142044f0025e388edc7ad7d11cfb3" Feb 25 14:21:56 crc kubenswrapper[4815]: I0225 14:21:56.402145 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98ad9145281ddac09a41d46dc7918222f69142044f0025e388edc7ad7d11cfb3"} err="failed to get container status \"98ad9145281ddac09a41d46dc7918222f69142044f0025e388edc7ad7d11cfb3\": rpc error: code = NotFound desc = could not find container \"98ad9145281ddac09a41d46dc7918222f69142044f0025e388edc7ad7d11cfb3\": container with ID starting with 98ad9145281ddac09a41d46dc7918222f69142044f0025e388edc7ad7d11cfb3 not found: ID does not exist" Feb 25 14:21:56 crc kubenswrapper[4815]: I0225 14:21:56.402165 4815 scope.go:117] "RemoveContainer" containerID="6c1bd96c0be4678656b17b82ae9cedb4da1d997e951d26c6355bbed515bf8d3d" Feb 25 14:21:56 crc kubenswrapper[4815]: E0225 14:21:56.402481 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c1bd96c0be4678656b17b82ae9cedb4da1d997e951d26c6355bbed515bf8d3d\": container with ID starting with 6c1bd96c0be4678656b17b82ae9cedb4da1d997e951d26c6355bbed515bf8d3d not found: ID does not exist" containerID="6c1bd96c0be4678656b17b82ae9cedb4da1d997e951d26c6355bbed515bf8d3d" Feb 25 14:21:56 crc kubenswrapper[4815]: I0225 14:21:56.402536 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c1bd96c0be4678656b17b82ae9cedb4da1d997e951d26c6355bbed515bf8d3d"} err="failed to get container status \"6c1bd96c0be4678656b17b82ae9cedb4da1d997e951d26c6355bbed515bf8d3d\": rpc error: code = NotFound desc = could not find container \"6c1bd96c0be4678656b17b82ae9cedb4da1d997e951d26c6355bbed515bf8d3d\": container with ID starting with 6c1bd96c0be4678656b17b82ae9cedb4da1d997e951d26c6355bbed515bf8d3d not found: ID does not exist" Feb 25 14:21:56 crc kubenswrapper[4815]: I0225 14:21:56.402553 4815 scope.go:117] "RemoveContainer" containerID="151bb90b4e14e3597759cc3ae141ba73af0fc25992b21d8beb3f12e019735140" Feb 25 14:21:56 crc kubenswrapper[4815]: E0225 14:21:56.402754 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"151bb90b4e14e3597759cc3ae141ba73af0fc25992b21d8beb3f12e019735140\": container with ID starting with 151bb90b4e14e3597759cc3ae141ba73af0fc25992b21d8beb3f12e019735140 not found: ID does not exist" containerID="151bb90b4e14e3597759cc3ae141ba73af0fc25992b21d8beb3f12e019735140" Feb 25 14:21:56 crc kubenswrapper[4815]: I0225 14:21:56.402775 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"151bb90b4e14e3597759cc3ae141ba73af0fc25992b21d8beb3f12e019735140"} err="failed to get container status \"151bb90b4e14e3597759cc3ae141ba73af0fc25992b21d8beb3f12e019735140\": rpc error: code = NotFound desc = could not find container \"151bb90b4e14e3597759cc3ae141ba73af0fc25992b21d8beb3f12e019735140\": container with ID starting with 151bb90b4e14e3597759cc3ae141ba73af0fc25992b21d8beb3f12e019735140 not found: ID does not exist" Feb 25 14:21:56 crc kubenswrapper[4815]: I0225 14:21:56.947351 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15dc1c43-85d2-44ee-92a5-7094b563550e" path="/var/lib/kubelet/pods/15dc1c43-85d2-44ee-92a5-7094b563550e/volumes" Feb 25 14:22:00 crc kubenswrapper[4815]: I0225 14:22:00.153361 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533822-2dhdp"] Feb 25 14:22:00 crc kubenswrapper[4815]: E0225 14:22:00.155656 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15dc1c43-85d2-44ee-92a5-7094b563550e" containerName="registry-server" Feb 25 14:22:00 crc kubenswrapper[4815]: I0225 14:22:00.155681 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="15dc1c43-85d2-44ee-92a5-7094b563550e" containerName="registry-server" Feb 25 14:22:00 crc kubenswrapper[4815]: E0225 14:22:00.155697 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15dc1c43-85d2-44ee-92a5-7094b563550e" containerName="extract-content" Feb 25 14:22:00 crc kubenswrapper[4815]: I0225 14:22:00.155706 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="15dc1c43-85d2-44ee-92a5-7094b563550e" containerName="extract-content" Feb 25 14:22:00 crc kubenswrapper[4815]: E0225 14:22:00.155748 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15dc1c43-85d2-44ee-92a5-7094b563550e" containerName="extract-utilities" Feb 25 14:22:00 crc kubenswrapper[4815]: I0225 14:22:00.155757 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="15dc1c43-85d2-44ee-92a5-7094b563550e" containerName="extract-utilities" Feb 25 14:22:00 crc kubenswrapper[4815]: I0225 14:22:00.156007 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="15dc1c43-85d2-44ee-92a5-7094b563550e" containerName="registry-server" Feb 25 14:22:00 crc kubenswrapper[4815]: I0225 14:22:00.156741 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533822-2dhdp" Feb 25 14:22:00 crc kubenswrapper[4815]: I0225 14:22:00.160531 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 14:22:00 crc kubenswrapper[4815]: I0225 14:22:00.160533 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 14:22:00 crc kubenswrapper[4815]: I0225 14:22:00.160852 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 14:22:00 crc kubenswrapper[4815]: I0225 14:22:00.168809 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533822-2dhdp"] Feb 25 14:22:00 crc kubenswrapper[4815]: I0225 14:22:00.352935 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvgz6\" (UniqueName: \"kubernetes.io/projected/90c81a61-15db-4959-8b68-580f2145bda9-kube-api-access-rvgz6\") pod \"auto-csr-approver-29533822-2dhdp\" (UID: \"90c81a61-15db-4959-8b68-580f2145bda9\") " pod="openshift-infra/auto-csr-approver-29533822-2dhdp" Feb 25 14:22:00 crc kubenswrapper[4815]: I0225 14:22:00.455041 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvgz6\" (UniqueName: \"kubernetes.io/projected/90c81a61-15db-4959-8b68-580f2145bda9-kube-api-access-rvgz6\") pod \"auto-csr-approver-29533822-2dhdp\" (UID: \"90c81a61-15db-4959-8b68-580f2145bda9\") " pod="openshift-infra/auto-csr-approver-29533822-2dhdp" Feb 25 14:22:00 crc kubenswrapper[4815]: I0225 14:22:00.473529 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvgz6\" (UniqueName: \"kubernetes.io/projected/90c81a61-15db-4959-8b68-580f2145bda9-kube-api-access-rvgz6\") pod \"auto-csr-approver-29533822-2dhdp\" (UID: \"90c81a61-15db-4959-8b68-580f2145bda9\") " pod="openshift-infra/auto-csr-approver-29533822-2dhdp" Feb 25 14:22:00 crc kubenswrapper[4815]: I0225 14:22:00.477593 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533822-2dhdp" Feb 25 14:22:00 crc kubenswrapper[4815]: I0225 14:22:00.889913 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533822-2dhdp"] Feb 25 14:22:01 crc kubenswrapper[4815]: I0225 14:22:01.337519 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533822-2dhdp" event={"ID":"90c81a61-15db-4959-8b68-580f2145bda9","Type":"ContainerStarted","Data":"7c2d259367491f5b7192a34045e5093c1035c8f33cb8e095b3f722b27b28b1a3"} Feb 25 14:22:02 crc kubenswrapper[4815]: I0225 14:22:02.349275 4815 generic.go:334] "Generic (PLEG): container finished" podID="90c81a61-15db-4959-8b68-580f2145bda9" containerID="737655317b1d56496d1b9d61c51583e44f57b9ca407d6954775524d7c8cdd9a5" exitCode=0 Feb 25 14:22:02 crc kubenswrapper[4815]: I0225 14:22:02.349336 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533822-2dhdp" event={"ID":"90c81a61-15db-4959-8b68-580f2145bda9","Type":"ContainerDied","Data":"737655317b1d56496d1b9d61c51583e44f57b9ca407d6954775524d7c8cdd9a5"} Feb 25 14:22:03 crc kubenswrapper[4815]: I0225 14:22:03.734383 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533822-2dhdp" Feb 25 14:22:03 crc kubenswrapper[4815]: I0225 14:22:03.923974 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvgz6\" (UniqueName: \"kubernetes.io/projected/90c81a61-15db-4959-8b68-580f2145bda9-kube-api-access-rvgz6\") pod \"90c81a61-15db-4959-8b68-580f2145bda9\" (UID: \"90c81a61-15db-4959-8b68-580f2145bda9\") " Feb 25 14:22:03 crc kubenswrapper[4815]: I0225 14:22:03.932303 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90c81a61-15db-4959-8b68-580f2145bda9-kube-api-access-rvgz6" (OuterVolumeSpecName: "kube-api-access-rvgz6") pod "90c81a61-15db-4959-8b68-580f2145bda9" (UID: "90c81a61-15db-4959-8b68-580f2145bda9"). InnerVolumeSpecName "kube-api-access-rvgz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:22:04 crc kubenswrapper[4815]: I0225 14:22:04.026740 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvgz6\" (UniqueName: \"kubernetes.io/projected/90c81a61-15db-4959-8b68-580f2145bda9-kube-api-access-rvgz6\") on node \"crc\" DevicePath \"\"" Feb 25 14:22:04 crc kubenswrapper[4815]: I0225 14:22:04.373638 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533822-2dhdp" event={"ID":"90c81a61-15db-4959-8b68-580f2145bda9","Type":"ContainerDied","Data":"7c2d259367491f5b7192a34045e5093c1035c8f33cb8e095b3f722b27b28b1a3"} Feb 25 14:22:04 crc kubenswrapper[4815]: I0225 14:22:04.373673 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c2d259367491f5b7192a34045e5093c1035c8f33cb8e095b3f722b27b28b1a3" Feb 25 14:22:04 crc kubenswrapper[4815]: I0225 14:22:04.373785 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533822-2dhdp" Feb 25 14:22:04 crc kubenswrapper[4815]: I0225 14:22:04.816872 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533816-pj8lb"] Feb 25 14:22:04 crc kubenswrapper[4815]: I0225 14:22:04.827397 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533816-pj8lb"] Feb 25 14:22:04 crc kubenswrapper[4815]: I0225 14:22:04.952936 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1" path="/var/lib/kubelet/pods/5a07e5f8-f7b6-4b8f-a7f6-1f99b90043a1/volumes" Feb 25 14:22:05 crc kubenswrapper[4815]: I0225 14:22:05.935875 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:22:05 crc kubenswrapper[4815]: E0225 14:22:05.937653 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:22:08 crc kubenswrapper[4815]: I0225 14:22:08.991543 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fds99"] Feb 25 14:22:08 crc kubenswrapper[4815]: E0225 14:22:08.992499 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90c81a61-15db-4959-8b68-580f2145bda9" containerName="oc" Feb 25 14:22:08 crc kubenswrapper[4815]: I0225 14:22:08.992607 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="90c81a61-15db-4959-8b68-580f2145bda9" containerName="oc" Feb 25 14:22:08 crc kubenswrapper[4815]: I0225 14:22:08.992898 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="90c81a61-15db-4959-8b68-580f2145bda9" containerName="oc" Feb 25 14:22:08 crc kubenswrapper[4815]: I0225 14:22:08.994731 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fds99" Feb 25 14:22:09 crc kubenswrapper[4815]: I0225 14:22:09.001235 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fds99"] Feb 25 14:22:09 crc kubenswrapper[4815]: I0225 14:22:09.139602 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-utilities\") pod \"community-operators-fds99\" (UID: \"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc\") " pod="openshift-marketplace/community-operators-fds99" Feb 25 14:22:09 crc kubenswrapper[4815]: I0225 14:22:09.140104 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9hwc\" (UniqueName: \"kubernetes.io/projected/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-kube-api-access-c9hwc\") pod \"community-operators-fds99\" (UID: \"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc\") " pod="openshift-marketplace/community-operators-fds99" Feb 25 14:22:09 crc kubenswrapper[4815]: I0225 14:22:09.140200 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-catalog-content\") pod \"community-operators-fds99\" (UID: \"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc\") " pod="openshift-marketplace/community-operators-fds99" Feb 25 14:22:09 crc kubenswrapper[4815]: I0225 14:22:09.241350 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-utilities\") pod \"community-operators-fds99\" (UID: \"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc\") " pod="openshift-marketplace/community-operators-fds99" Feb 25 14:22:09 crc kubenswrapper[4815]: I0225 14:22:09.241440 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9hwc\" (UniqueName: \"kubernetes.io/projected/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-kube-api-access-c9hwc\") pod \"community-operators-fds99\" (UID: \"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc\") " pod="openshift-marketplace/community-operators-fds99" Feb 25 14:22:09 crc kubenswrapper[4815]: I0225 14:22:09.241536 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-catalog-content\") pod \"community-operators-fds99\" (UID: \"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc\") " pod="openshift-marketplace/community-operators-fds99" Feb 25 14:22:09 crc kubenswrapper[4815]: I0225 14:22:09.241901 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-utilities\") pod \"community-operators-fds99\" (UID: \"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc\") " pod="openshift-marketplace/community-operators-fds99" Feb 25 14:22:09 crc kubenswrapper[4815]: I0225 14:22:09.241908 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-catalog-content\") pod \"community-operators-fds99\" (UID: \"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc\") " pod="openshift-marketplace/community-operators-fds99" Feb 25 14:22:09 crc kubenswrapper[4815]: I0225 14:22:09.264501 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9hwc\" (UniqueName: \"kubernetes.io/projected/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-kube-api-access-c9hwc\") pod \"community-operators-fds99\" (UID: \"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc\") " pod="openshift-marketplace/community-operators-fds99" Feb 25 14:22:09 crc kubenswrapper[4815]: I0225 14:22:09.329710 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fds99" Feb 25 14:22:09 crc kubenswrapper[4815]: I0225 14:22:09.859034 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fds99"] Feb 25 14:22:10 crc kubenswrapper[4815]: I0225 14:22:10.434356 4815 generic.go:334] "Generic (PLEG): container finished" podID="10ec0fa0-1cf2-4264-8e3a-55f2e13193dc" containerID="05f3cceb164dbed6c2990e902762ab88a7db8bac4444a065ee42f4793c7be732" exitCode=0 Feb 25 14:22:10 crc kubenswrapper[4815]: I0225 14:22:10.434550 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fds99" event={"ID":"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc","Type":"ContainerDied","Data":"05f3cceb164dbed6c2990e902762ab88a7db8bac4444a065ee42f4793c7be732"} Feb 25 14:22:10 crc kubenswrapper[4815]: I0225 14:22:10.434647 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fds99" event={"ID":"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc","Type":"ContainerStarted","Data":"02ca3d61a0b80af272b243855277b6b0c363af3d68996eccb4d2988a762adbcf"} Feb 25 14:22:11 crc kubenswrapper[4815]: I0225 14:22:11.447620 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fds99" event={"ID":"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc","Type":"ContainerStarted","Data":"1d60d454e5e1ece03bc79a188f86f7ed43ab18b01bdf661cc1620ef5fd803b87"} Feb 25 14:22:13 crc kubenswrapper[4815]: I0225 14:22:13.467895 4815 generic.go:334] "Generic (PLEG): container finished" podID="10ec0fa0-1cf2-4264-8e3a-55f2e13193dc" containerID="1d60d454e5e1ece03bc79a188f86f7ed43ab18b01bdf661cc1620ef5fd803b87" exitCode=0 Feb 25 14:22:13 crc kubenswrapper[4815]: I0225 14:22:13.467998 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fds99" event={"ID":"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc","Type":"ContainerDied","Data":"1d60d454e5e1ece03bc79a188f86f7ed43ab18b01bdf661cc1620ef5fd803b87"} Feb 25 14:22:14 crc kubenswrapper[4815]: I0225 14:22:14.482820 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fds99" event={"ID":"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc","Type":"ContainerStarted","Data":"55433acd1ef3a868be66fd65badf2d33ec9ea53f4f7ca5bc3b89de602308d1ab"} Feb 25 14:22:14 crc kubenswrapper[4815]: I0225 14:22:14.512823 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fds99" podStartSLOduration=3.016033181 podStartE2EDuration="6.512801838s" podCreationTimestamp="2026-02-25 14:22:08 +0000 UTC" firstStartedPulling="2026-02-25 14:22:10.436637638 +0000 UTC m=+3688.237735742" lastFinishedPulling="2026-02-25 14:22:13.933406345 +0000 UTC m=+3691.734504399" observedRunningTime="2026-02-25 14:22:14.507883285 +0000 UTC m=+3692.308981359" watchObservedRunningTime="2026-02-25 14:22:14.512801838 +0000 UTC m=+3692.313899892" Feb 25 14:22:19 crc kubenswrapper[4815]: I0225 14:22:19.330500 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fds99" Feb 25 14:22:19 crc kubenswrapper[4815]: I0225 14:22:19.332271 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fds99" Feb 25 14:22:19 crc kubenswrapper[4815]: I0225 14:22:19.393984 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fds99" Feb 25 14:22:19 crc kubenswrapper[4815]: I0225 14:22:19.582116 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fds99" Feb 25 14:22:19 crc kubenswrapper[4815]: I0225 14:22:19.648844 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fds99"] Feb 25 14:22:20 crc kubenswrapper[4815]: I0225 14:22:20.935790 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:22:20 crc kubenswrapper[4815]: E0225 14:22:20.936081 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:22:21 crc kubenswrapper[4815]: I0225 14:22:21.555477 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fds99" podUID="10ec0fa0-1cf2-4264-8e3a-55f2e13193dc" containerName="registry-server" containerID="cri-o://55433acd1ef3a868be66fd65badf2d33ec9ea53f4f7ca5bc3b89de602308d1ab" gracePeriod=2 Feb 25 14:22:21 crc kubenswrapper[4815]: I0225 14:22:21.972459 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fds99" Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.116641 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-catalog-content\") pod \"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc\" (UID: \"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc\") " Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.116788 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9hwc\" (UniqueName: \"kubernetes.io/projected/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-kube-api-access-c9hwc\") pod \"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc\" (UID: \"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc\") " Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.116961 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-utilities\") pod \"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc\" (UID: \"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc\") " Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.118281 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-utilities" (OuterVolumeSpecName: "utilities") pod "10ec0fa0-1cf2-4264-8e3a-55f2e13193dc" (UID: "10ec0fa0-1cf2-4264-8e3a-55f2e13193dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.126840 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-kube-api-access-c9hwc" (OuterVolumeSpecName: "kube-api-access-c9hwc") pod "10ec0fa0-1cf2-4264-8e3a-55f2e13193dc" (UID: "10ec0fa0-1cf2-4264-8e3a-55f2e13193dc"). InnerVolumeSpecName "kube-api-access-c9hwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.219771 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9hwc\" (UniqueName: \"kubernetes.io/projected/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-kube-api-access-c9hwc\") on node \"crc\" DevicePath \"\"" Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.219821 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.518845 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10ec0fa0-1cf2-4264-8e3a-55f2e13193dc" (UID: "10ec0fa0-1cf2-4264-8e3a-55f2e13193dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.526126 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.567165 4815 generic.go:334] "Generic (PLEG): container finished" podID="10ec0fa0-1cf2-4264-8e3a-55f2e13193dc" containerID="55433acd1ef3a868be66fd65badf2d33ec9ea53f4f7ca5bc3b89de602308d1ab" exitCode=0 Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.567215 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fds99" event={"ID":"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc","Type":"ContainerDied","Data":"55433acd1ef3a868be66fd65badf2d33ec9ea53f4f7ca5bc3b89de602308d1ab"} Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.567235 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fds99" Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.567253 4815 scope.go:117] "RemoveContainer" containerID="55433acd1ef3a868be66fd65badf2d33ec9ea53f4f7ca5bc3b89de602308d1ab" Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.567242 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fds99" event={"ID":"10ec0fa0-1cf2-4264-8e3a-55f2e13193dc","Type":"ContainerDied","Data":"02ca3d61a0b80af272b243855277b6b0c363af3d68996eccb4d2988a762adbcf"} Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.593497 4815 scope.go:117] "RemoveContainer" containerID="1d60d454e5e1ece03bc79a188f86f7ed43ab18b01bdf661cc1620ef5fd803b87" Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.626464 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fds99"] Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.628939 4815 scope.go:117] "RemoveContainer" containerID="05f3cceb164dbed6c2990e902762ab88a7db8bac4444a065ee42f4793c7be732" Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.634002 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fds99"] Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.674082 4815 scope.go:117] "RemoveContainer" containerID="55433acd1ef3a868be66fd65badf2d33ec9ea53f4f7ca5bc3b89de602308d1ab" Feb 25 14:22:22 crc kubenswrapper[4815]: E0225 14:22:22.674639 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55433acd1ef3a868be66fd65badf2d33ec9ea53f4f7ca5bc3b89de602308d1ab\": container with ID starting with 55433acd1ef3a868be66fd65badf2d33ec9ea53f4f7ca5bc3b89de602308d1ab not found: ID does not exist" containerID="55433acd1ef3a868be66fd65badf2d33ec9ea53f4f7ca5bc3b89de602308d1ab" Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.674680 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55433acd1ef3a868be66fd65badf2d33ec9ea53f4f7ca5bc3b89de602308d1ab"} err="failed to get container status \"55433acd1ef3a868be66fd65badf2d33ec9ea53f4f7ca5bc3b89de602308d1ab\": rpc error: code = NotFound desc = could not find container \"55433acd1ef3a868be66fd65badf2d33ec9ea53f4f7ca5bc3b89de602308d1ab\": container with ID starting with 55433acd1ef3a868be66fd65badf2d33ec9ea53f4f7ca5bc3b89de602308d1ab not found: ID does not exist" Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.674704 4815 scope.go:117] "RemoveContainer" containerID="1d60d454e5e1ece03bc79a188f86f7ed43ab18b01bdf661cc1620ef5fd803b87" Feb 25 14:22:22 crc kubenswrapper[4815]: E0225 14:22:22.675069 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d60d454e5e1ece03bc79a188f86f7ed43ab18b01bdf661cc1620ef5fd803b87\": container with ID starting with 1d60d454e5e1ece03bc79a188f86f7ed43ab18b01bdf661cc1620ef5fd803b87 not found: ID does not exist" containerID="1d60d454e5e1ece03bc79a188f86f7ed43ab18b01bdf661cc1620ef5fd803b87" Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.675098 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d60d454e5e1ece03bc79a188f86f7ed43ab18b01bdf661cc1620ef5fd803b87"} err="failed to get container status \"1d60d454e5e1ece03bc79a188f86f7ed43ab18b01bdf661cc1620ef5fd803b87\": rpc error: code = NotFound desc = could not find container \"1d60d454e5e1ece03bc79a188f86f7ed43ab18b01bdf661cc1620ef5fd803b87\": container with ID starting with 1d60d454e5e1ece03bc79a188f86f7ed43ab18b01bdf661cc1620ef5fd803b87 not found: ID does not exist" Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.675114 4815 scope.go:117] "RemoveContainer" containerID="05f3cceb164dbed6c2990e902762ab88a7db8bac4444a065ee42f4793c7be732" Feb 25 14:22:22 crc kubenswrapper[4815]: E0225 14:22:22.675483 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05f3cceb164dbed6c2990e902762ab88a7db8bac4444a065ee42f4793c7be732\": container with ID starting with 05f3cceb164dbed6c2990e902762ab88a7db8bac4444a065ee42f4793c7be732 not found: ID does not exist" containerID="05f3cceb164dbed6c2990e902762ab88a7db8bac4444a065ee42f4793c7be732" Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.675525 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05f3cceb164dbed6c2990e902762ab88a7db8bac4444a065ee42f4793c7be732"} err="failed to get container status \"05f3cceb164dbed6c2990e902762ab88a7db8bac4444a065ee42f4793c7be732\": rpc error: code = NotFound desc = could not find container \"05f3cceb164dbed6c2990e902762ab88a7db8bac4444a065ee42f4793c7be732\": container with ID starting with 05f3cceb164dbed6c2990e902762ab88a7db8bac4444a065ee42f4793c7be732 not found: ID does not exist" Feb 25 14:22:22 crc kubenswrapper[4815]: I0225 14:22:22.950994 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10ec0fa0-1cf2-4264-8e3a-55f2e13193dc" path="/var/lib/kubelet/pods/10ec0fa0-1cf2-4264-8e3a-55f2e13193dc/volumes" Feb 25 14:22:34 crc kubenswrapper[4815]: I0225 14:22:34.936433 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:22:34 crc kubenswrapper[4815]: E0225 14:22:34.937862 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:22:45 crc kubenswrapper[4815]: I0225 14:22:45.935610 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:22:45 crc kubenswrapper[4815]: E0225 14:22:45.936443 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:22:50 crc kubenswrapper[4815]: I0225 14:22:50.205041 4815 scope.go:117] "RemoveContainer" containerID="2b8f0cbbadd1b0bb08e8907e0d677eef59504fdca28280b5bf3a5f296918cfb1" Feb 25 14:22:50 crc kubenswrapper[4815]: I0225 14:22:50.228069 4815 scope.go:117] "RemoveContainer" containerID="1c9a31b8117a83d4c038cab9f9f6330dba60646d171fc2e6dba06da20c2a0966" Feb 25 14:22:57 crc kubenswrapper[4815]: I0225 14:22:57.941012 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:22:57 crc kubenswrapper[4815]: E0225 14:22:57.951029 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:23:10 crc kubenswrapper[4815]: I0225 14:23:10.936901 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:23:10 crc kubenswrapper[4815]: E0225 14:23:10.938481 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:23:25 crc kubenswrapper[4815]: I0225 14:23:25.935991 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:23:25 crc kubenswrapper[4815]: E0225 14:23:25.937059 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:23:39 crc kubenswrapper[4815]: I0225 14:23:39.936004 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:23:39 crc kubenswrapper[4815]: E0225 14:23:39.936711 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:23:51 crc kubenswrapper[4815]: I0225 14:23:51.936051 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:23:51 crc kubenswrapper[4815]: E0225 14:23:51.936872 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:24:00 crc kubenswrapper[4815]: I0225 14:24:00.166010 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533824-wqg2c"] Feb 25 14:24:00 crc kubenswrapper[4815]: E0225 14:24:00.167350 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10ec0fa0-1cf2-4264-8e3a-55f2e13193dc" containerName="extract-utilities" Feb 25 14:24:00 crc kubenswrapper[4815]: I0225 14:24:00.167374 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="10ec0fa0-1cf2-4264-8e3a-55f2e13193dc" containerName="extract-utilities" Feb 25 14:24:00 crc kubenswrapper[4815]: E0225 14:24:00.167464 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10ec0fa0-1cf2-4264-8e3a-55f2e13193dc" containerName="registry-server" Feb 25 14:24:00 crc kubenswrapper[4815]: I0225 14:24:00.167480 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="10ec0fa0-1cf2-4264-8e3a-55f2e13193dc" containerName="registry-server" Feb 25 14:24:00 crc kubenswrapper[4815]: E0225 14:24:00.167493 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10ec0fa0-1cf2-4264-8e3a-55f2e13193dc" containerName="extract-content" Feb 25 14:24:00 crc kubenswrapper[4815]: I0225 14:24:00.167504 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="10ec0fa0-1cf2-4264-8e3a-55f2e13193dc" containerName="extract-content" Feb 25 14:24:00 crc kubenswrapper[4815]: I0225 14:24:00.167889 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="10ec0fa0-1cf2-4264-8e3a-55f2e13193dc" containerName="registry-server" Feb 25 14:24:00 crc kubenswrapper[4815]: I0225 14:24:00.168752 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533824-wqg2c"] Feb 25 14:24:00 crc kubenswrapper[4815]: I0225 14:24:00.168866 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533824-wqg2c" Feb 25 14:24:00 crc kubenswrapper[4815]: I0225 14:24:00.193256 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 14:24:00 crc kubenswrapper[4815]: I0225 14:24:00.193497 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 14:24:00 crc kubenswrapper[4815]: I0225 14:24:00.194210 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 14:24:00 crc kubenswrapper[4815]: I0225 14:24:00.331157 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvs2h\" (UniqueName: \"kubernetes.io/projected/8620949a-fbdc-402c-8597-7aa16679d353-kube-api-access-wvs2h\") pod \"auto-csr-approver-29533824-wqg2c\" (UID: \"8620949a-fbdc-402c-8597-7aa16679d353\") " pod="openshift-infra/auto-csr-approver-29533824-wqg2c" Feb 25 14:24:00 crc kubenswrapper[4815]: I0225 14:24:00.433032 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvs2h\" (UniqueName: \"kubernetes.io/projected/8620949a-fbdc-402c-8597-7aa16679d353-kube-api-access-wvs2h\") pod \"auto-csr-approver-29533824-wqg2c\" (UID: \"8620949a-fbdc-402c-8597-7aa16679d353\") " pod="openshift-infra/auto-csr-approver-29533824-wqg2c" Feb 25 14:24:00 crc kubenswrapper[4815]: I0225 14:24:00.457312 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvs2h\" (UniqueName: \"kubernetes.io/projected/8620949a-fbdc-402c-8597-7aa16679d353-kube-api-access-wvs2h\") pod \"auto-csr-approver-29533824-wqg2c\" (UID: \"8620949a-fbdc-402c-8597-7aa16679d353\") " pod="openshift-infra/auto-csr-approver-29533824-wqg2c" Feb 25 14:24:00 crc kubenswrapper[4815]: I0225 14:24:00.508956 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533824-wqg2c" Feb 25 14:24:00 crc kubenswrapper[4815]: I0225 14:24:00.969040 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533824-wqg2c"] Feb 25 14:24:00 crc kubenswrapper[4815]: I0225 14:24:00.970661 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 14:24:01 crc kubenswrapper[4815]: I0225 14:24:01.536891 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533824-wqg2c" event={"ID":"8620949a-fbdc-402c-8597-7aa16679d353","Type":"ContainerStarted","Data":"4e90b0d650420839bfa7a3e4b4e86a0072eb174fd63ee24ebd2dd69df904e291"} Feb 25 14:24:02 crc kubenswrapper[4815]: I0225 14:24:02.548977 4815 generic.go:334] "Generic (PLEG): container finished" podID="8620949a-fbdc-402c-8597-7aa16679d353" containerID="16f28b33e82753de5c53b609b4f31e3e3a4cbc8ae45913e7504cb677bd429f6b" exitCode=0 Feb 25 14:24:02 crc kubenswrapper[4815]: I0225 14:24:02.549018 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533824-wqg2c" event={"ID":"8620949a-fbdc-402c-8597-7aa16679d353","Type":"ContainerDied","Data":"16f28b33e82753de5c53b609b4f31e3e3a4cbc8ae45913e7504cb677bd429f6b"} Feb 25 14:24:02 crc kubenswrapper[4815]: I0225 14:24:02.960372 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:24:02 crc kubenswrapper[4815]: E0225 14:24:02.961008 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:24:03 crc kubenswrapper[4815]: I0225 14:24:03.909166 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533824-wqg2c" Feb 25 14:24:03 crc kubenswrapper[4815]: I0225 14:24:03.999538 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvs2h\" (UniqueName: \"kubernetes.io/projected/8620949a-fbdc-402c-8597-7aa16679d353-kube-api-access-wvs2h\") pod \"8620949a-fbdc-402c-8597-7aa16679d353\" (UID: \"8620949a-fbdc-402c-8597-7aa16679d353\") " Feb 25 14:24:04 crc kubenswrapper[4815]: I0225 14:24:04.006716 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8620949a-fbdc-402c-8597-7aa16679d353-kube-api-access-wvs2h" (OuterVolumeSpecName: "kube-api-access-wvs2h") pod "8620949a-fbdc-402c-8597-7aa16679d353" (UID: "8620949a-fbdc-402c-8597-7aa16679d353"). InnerVolumeSpecName "kube-api-access-wvs2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:24:04 crc kubenswrapper[4815]: I0225 14:24:04.103460 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvs2h\" (UniqueName: \"kubernetes.io/projected/8620949a-fbdc-402c-8597-7aa16679d353-kube-api-access-wvs2h\") on node \"crc\" DevicePath \"\"" Feb 25 14:24:04 crc kubenswrapper[4815]: I0225 14:24:04.565538 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533824-wqg2c" event={"ID":"8620949a-fbdc-402c-8597-7aa16679d353","Type":"ContainerDied","Data":"4e90b0d650420839bfa7a3e4b4e86a0072eb174fd63ee24ebd2dd69df904e291"} Feb 25 14:24:04 crc kubenswrapper[4815]: I0225 14:24:04.565809 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e90b0d650420839bfa7a3e4b4e86a0072eb174fd63ee24ebd2dd69df904e291" Feb 25 14:24:04 crc kubenswrapper[4815]: I0225 14:24:04.565580 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533824-wqg2c" Feb 25 14:24:05 crc kubenswrapper[4815]: I0225 14:24:05.006783 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533818-2wncd"] Feb 25 14:24:05 crc kubenswrapper[4815]: I0225 14:24:05.014116 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533818-2wncd"] Feb 25 14:24:06 crc kubenswrapper[4815]: I0225 14:24:06.949306 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23253ded-9aae-451f-a7fc-1b10518f4312" path="/var/lib/kubelet/pods/23253ded-9aae-451f-a7fc-1b10518f4312/volumes" Feb 25 14:24:17 crc kubenswrapper[4815]: I0225 14:24:17.936165 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:24:17 crc kubenswrapper[4815]: E0225 14:24:17.938638 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:24:29 crc kubenswrapper[4815]: I0225 14:24:29.936206 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:24:29 crc kubenswrapper[4815]: E0225 14:24:29.936999 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:24:43 crc kubenswrapper[4815]: I0225 14:24:43.936072 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:24:43 crc kubenswrapper[4815]: E0225 14:24:43.936892 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:24:50 crc kubenswrapper[4815]: I0225 14:24:50.385155 4815 scope.go:117] "RemoveContainer" containerID="6b18ca2315823c26b460d836ee0e1a4578448d2af942774f6b847a88525d1622" Feb 25 14:24:57 crc kubenswrapper[4815]: I0225 14:24:57.936877 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:24:57 crc kubenswrapper[4815]: E0225 14:24:57.938263 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:25:10 crc kubenswrapper[4815]: I0225 14:25:10.941141 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:25:10 crc kubenswrapper[4815]: E0225 14:25:10.943229 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:25:21 crc kubenswrapper[4815]: I0225 14:25:21.936029 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:25:21 crc kubenswrapper[4815]: E0225 14:25:21.936927 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:25:26 crc kubenswrapper[4815]: I0225 14:25:26.912980 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7p5pw"] Feb 25 14:25:26 crc kubenswrapper[4815]: E0225 14:25:26.915648 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8620949a-fbdc-402c-8597-7aa16679d353" containerName="oc" Feb 25 14:25:26 crc kubenswrapper[4815]: I0225 14:25:26.915819 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8620949a-fbdc-402c-8597-7aa16679d353" containerName="oc" Feb 25 14:25:26 crc kubenswrapper[4815]: I0225 14:25:26.916310 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8620949a-fbdc-402c-8597-7aa16679d353" containerName="oc" Feb 25 14:25:26 crc kubenswrapper[4815]: I0225 14:25:26.918562 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7p5pw" Feb 25 14:25:26 crc kubenswrapper[4815]: I0225 14:25:26.930396 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7p5pw"] Feb 25 14:25:26 crc kubenswrapper[4815]: I0225 14:25:26.991576 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84b81897-b136-4ce2-9c5a-6f3bae225e48-utilities\") pod \"redhat-marketplace-7p5pw\" (UID: \"84b81897-b136-4ce2-9c5a-6f3bae225e48\") " pod="openshift-marketplace/redhat-marketplace-7p5pw" Feb 25 14:25:26 crc kubenswrapper[4815]: I0225 14:25:26.991650 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7lsp\" (UniqueName: \"kubernetes.io/projected/84b81897-b136-4ce2-9c5a-6f3bae225e48-kube-api-access-z7lsp\") pod \"redhat-marketplace-7p5pw\" (UID: \"84b81897-b136-4ce2-9c5a-6f3bae225e48\") " pod="openshift-marketplace/redhat-marketplace-7p5pw" Feb 25 14:25:26 crc kubenswrapper[4815]: I0225 14:25:26.991784 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84b81897-b136-4ce2-9c5a-6f3bae225e48-catalog-content\") pod \"redhat-marketplace-7p5pw\" (UID: \"84b81897-b136-4ce2-9c5a-6f3bae225e48\") " pod="openshift-marketplace/redhat-marketplace-7p5pw" Feb 25 14:25:27 crc kubenswrapper[4815]: I0225 14:25:27.093405 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84b81897-b136-4ce2-9c5a-6f3bae225e48-utilities\") pod \"redhat-marketplace-7p5pw\" (UID: \"84b81897-b136-4ce2-9c5a-6f3bae225e48\") " pod="openshift-marketplace/redhat-marketplace-7p5pw" Feb 25 14:25:27 crc kubenswrapper[4815]: I0225 14:25:27.093470 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7lsp\" (UniqueName: \"kubernetes.io/projected/84b81897-b136-4ce2-9c5a-6f3bae225e48-kube-api-access-z7lsp\") pod \"redhat-marketplace-7p5pw\" (UID: \"84b81897-b136-4ce2-9c5a-6f3bae225e48\") " pod="openshift-marketplace/redhat-marketplace-7p5pw" Feb 25 14:25:27 crc kubenswrapper[4815]: I0225 14:25:27.093549 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84b81897-b136-4ce2-9c5a-6f3bae225e48-catalog-content\") pod \"redhat-marketplace-7p5pw\" (UID: \"84b81897-b136-4ce2-9c5a-6f3bae225e48\") " pod="openshift-marketplace/redhat-marketplace-7p5pw" Feb 25 14:25:27 crc kubenswrapper[4815]: I0225 14:25:27.093995 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84b81897-b136-4ce2-9c5a-6f3bae225e48-catalog-content\") pod \"redhat-marketplace-7p5pw\" (UID: \"84b81897-b136-4ce2-9c5a-6f3bae225e48\") " pod="openshift-marketplace/redhat-marketplace-7p5pw" Feb 25 14:25:27 crc kubenswrapper[4815]: I0225 14:25:27.094178 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84b81897-b136-4ce2-9c5a-6f3bae225e48-utilities\") pod \"redhat-marketplace-7p5pw\" (UID: \"84b81897-b136-4ce2-9c5a-6f3bae225e48\") " pod="openshift-marketplace/redhat-marketplace-7p5pw" Feb 25 14:25:27 crc kubenswrapper[4815]: I0225 14:25:27.114459 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7lsp\" (UniqueName: \"kubernetes.io/projected/84b81897-b136-4ce2-9c5a-6f3bae225e48-kube-api-access-z7lsp\") pod \"redhat-marketplace-7p5pw\" (UID: \"84b81897-b136-4ce2-9c5a-6f3bae225e48\") " pod="openshift-marketplace/redhat-marketplace-7p5pw" Feb 25 14:25:27 crc kubenswrapper[4815]: I0225 14:25:27.256809 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7p5pw" Feb 25 14:25:27 crc kubenswrapper[4815]: I0225 14:25:27.781609 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7p5pw"] Feb 25 14:25:28 crc kubenswrapper[4815]: I0225 14:25:28.347652 4815 generic.go:334] "Generic (PLEG): container finished" podID="84b81897-b136-4ce2-9c5a-6f3bae225e48" containerID="33f27c7f2ff401aacb9573d7e3729ac7d2a6b530d98267f5546daf4606c05b3a" exitCode=0 Feb 25 14:25:28 crc kubenswrapper[4815]: I0225 14:25:28.347746 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7p5pw" event={"ID":"84b81897-b136-4ce2-9c5a-6f3bae225e48","Type":"ContainerDied","Data":"33f27c7f2ff401aacb9573d7e3729ac7d2a6b530d98267f5546daf4606c05b3a"} Feb 25 14:25:28 crc kubenswrapper[4815]: I0225 14:25:28.347960 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7p5pw" event={"ID":"84b81897-b136-4ce2-9c5a-6f3bae225e48","Type":"ContainerStarted","Data":"6d48d94a89da1f7cac3baba8a23d69d09ef9e937f18a1536c834ed34e30676c4"} Feb 25 14:25:29 crc kubenswrapper[4815]: I0225 14:25:29.364311 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7p5pw" event={"ID":"84b81897-b136-4ce2-9c5a-6f3bae225e48","Type":"ContainerStarted","Data":"68e2db4ce18cc985aaebf39abbb39400317fa19f4f710971e73bcab5805ec9a9"} Feb 25 14:25:30 crc kubenswrapper[4815]: I0225 14:25:30.373560 4815 generic.go:334] "Generic (PLEG): container finished" podID="84b81897-b136-4ce2-9c5a-6f3bae225e48" containerID="68e2db4ce18cc985aaebf39abbb39400317fa19f4f710971e73bcab5805ec9a9" exitCode=0 Feb 25 14:25:30 crc kubenswrapper[4815]: I0225 14:25:30.373609 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7p5pw" event={"ID":"84b81897-b136-4ce2-9c5a-6f3bae225e48","Type":"ContainerDied","Data":"68e2db4ce18cc985aaebf39abbb39400317fa19f4f710971e73bcab5805ec9a9"} Feb 25 14:25:31 crc kubenswrapper[4815]: I0225 14:25:31.383449 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7p5pw" event={"ID":"84b81897-b136-4ce2-9c5a-6f3bae225e48","Type":"ContainerStarted","Data":"ce5fb86be2ef3e10843722e133fc588bfc6dc3a9f70e9a09918cb393f2c32783"} Feb 25 14:25:31 crc kubenswrapper[4815]: I0225 14:25:31.409015 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7p5pw" podStartSLOduration=2.885195753 podStartE2EDuration="5.408994737s" podCreationTimestamp="2026-02-25 14:25:26 +0000 UTC" firstStartedPulling="2026-02-25 14:25:28.349170728 +0000 UTC m=+3886.150268782" lastFinishedPulling="2026-02-25 14:25:30.872969692 +0000 UTC m=+3888.674067766" observedRunningTime="2026-02-25 14:25:31.400139001 +0000 UTC m=+3889.201237075" watchObservedRunningTime="2026-02-25 14:25:31.408994737 +0000 UTC m=+3889.210092801" Feb 25 14:25:32 crc kubenswrapper[4815]: I0225 14:25:32.942610 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:25:32 crc kubenswrapper[4815]: E0225 14:25:32.943131 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:25:37 crc kubenswrapper[4815]: I0225 14:25:37.257438 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7p5pw" Feb 25 14:25:37 crc kubenswrapper[4815]: I0225 14:25:37.258381 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7p5pw" Feb 25 14:25:37 crc kubenswrapper[4815]: I0225 14:25:37.314678 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7p5pw" Feb 25 14:25:37 crc kubenswrapper[4815]: I0225 14:25:37.480561 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7p5pw" Feb 25 14:25:38 crc kubenswrapper[4815]: I0225 14:25:38.481939 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7p5pw"] Feb 25 14:25:40 crc kubenswrapper[4815]: I0225 14:25:40.461755 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7p5pw" podUID="84b81897-b136-4ce2-9c5a-6f3bae225e48" containerName="registry-server" containerID="cri-o://ce5fb86be2ef3e10843722e133fc588bfc6dc3a9f70e9a09918cb393f2c32783" gracePeriod=2 Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.029588 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7p5pw" Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.188763 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84b81897-b136-4ce2-9c5a-6f3bae225e48-utilities\") pod \"84b81897-b136-4ce2-9c5a-6f3bae225e48\" (UID: \"84b81897-b136-4ce2-9c5a-6f3bae225e48\") " Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.189232 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84b81897-b136-4ce2-9c5a-6f3bae225e48-catalog-content\") pod \"84b81897-b136-4ce2-9c5a-6f3bae225e48\" (UID: \"84b81897-b136-4ce2-9c5a-6f3bae225e48\") " Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.189344 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7lsp\" (UniqueName: \"kubernetes.io/projected/84b81897-b136-4ce2-9c5a-6f3bae225e48-kube-api-access-z7lsp\") pod \"84b81897-b136-4ce2-9c5a-6f3bae225e48\" (UID: \"84b81897-b136-4ce2-9c5a-6f3bae225e48\") " Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.189866 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84b81897-b136-4ce2-9c5a-6f3bae225e48-utilities" (OuterVolumeSpecName: "utilities") pod "84b81897-b136-4ce2-9c5a-6f3bae225e48" (UID: "84b81897-b136-4ce2-9c5a-6f3bae225e48"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.194612 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84b81897-b136-4ce2-9c5a-6f3bae225e48-kube-api-access-z7lsp" (OuterVolumeSpecName: "kube-api-access-z7lsp") pod "84b81897-b136-4ce2-9c5a-6f3bae225e48" (UID: "84b81897-b136-4ce2-9c5a-6f3bae225e48"). InnerVolumeSpecName "kube-api-access-z7lsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.217628 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84b81897-b136-4ce2-9c5a-6f3bae225e48-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84b81897-b136-4ce2-9c5a-6f3bae225e48" (UID: "84b81897-b136-4ce2-9c5a-6f3bae225e48"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.291945 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7lsp\" (UniqueName: \"kubernetes.io/projected/84b81897-b136-4ce2-9c5a-6f3bae225e48-kube-api-access-z7lsp\") on node \"crc\" DevicePath \"\"" Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.291986 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84b81897-b136-4ce2-9c5a-6f3bae225e48-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.291999 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84b81897-b136-4ce2-9c5a-6f3bae225e48-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.473830 4815 generic.go:334] "Generic (PLEG): container finished" podID="84b81897-b136-4ce2-9c5a-6f3bae225e48" containerID="ce5fb86be2ef3e10843722e133fc588bfc6dc3a9f70e9a09918cb393f2c32783" exitCode=0 Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.473897 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7p5pw" event={"ID":"84b81897-b136-4ce2-9c5a-6f3bae225e48","Type":"ContainerDied","Data":"ce5fb86be2ef3e10843722e133fc588bfc6dc3a9f70e9a09918cb393f2c32783"} Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.473949 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7p5pw" event={"ID":"84b81897-b136-4ce2-9c5a-6f3bae225e48","Type":"ContainerDied","Data":"6d48d94a89da1f7cac3baba8a23d69d09ef9e937f18a1536c834ed34e30676c4"} Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.473971 4815 scope.go:117] "RemoveContainer" containerID="ce5fb86be2ef3e10843722e133fc588bfc6dc3a9f70e9a09918cb393f2c32783" Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.474179 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7p5pw" Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.518206 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7p5pw"] Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.522780 4815 scope.go:117] "RemoveContainer" containerID="68e2db4ce18cc985aaebf39abbb39400317fa19f4f710971e73bcab5805ec9a9" Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.527798 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7p5pw"] Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.550841 4815 scope.go:117] "RemoveContainer" containerID="33f27c7f2ff401aacb9573d7e3729ac7d2a6b530d98267f5546daf4606c05b3a" Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.587207 4815 scope.go:117] "RemoveContainer" containerID="ce5fb86be2ef3e10843722e133fc588bfc6dc3a9f70e9a09918cb393f2c32783" Feb 25 14:25:41 crc kubenswrapper[4815]: E0225 14:25:41.587906 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce5fb86be2ef3e10843722e133fc588bfc6dc3a9f70e9a09918cb393f2c32783\": container with ID starting with ce5fb86be2ef3e10843722e133fc588bfc6dc3a9f70e9a09918cb393f2c32783 not found: ID does not exist" containerID="ce5fb86be2ef3e10843722e133fc588bfc6dc3a9f70e9a09918cb393f2c32783" Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.587954 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce5fb86be2ef3e10843722e133fc588bfc6dc3a9f70e9a09918cb393f2c32783"} err="failed to get container status \"ce5fb86be2ef3e10843722e133fc588bfc6dc3a9f70e9a09918cb393f2c32783\": rpc error: code = NotFound desc = could not find container \"ce5fb86be2ef3e10843722e133fc588bfc6dc3a9f70e9a09918cb393f2c32783\": container with ID starting with ce5fb86be2ef3e10843722e133fc588bfc6dc3a9f70e9a09918cb393f2c32783 not found: ID does not exist" Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.587981 4815 scope.go:117] "RemoveContainer" containerID="68e2db4ce18cc985aaebf39abbb39400317fa19f4f710971e73bcab5805ec9a9" Feb 25 14:25:41 crc kubenswrapper[4815]: E0225 14:25:41.588216 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68e2db4ce18cc985aaebf39abbb39400317fa19f4f710971e73bcab5805ec9a9\": container with ID starting with 68e2db4ce18cc985aaebf39abbb39400317fa19f4f710971e73bcab5805ec9a9 not found: ID does not exist" containerID="68e2db4ce18cc985aaebf39abbb39400317fa19f4f710971e73bcab5805ec9a9" Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.588242 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68e2db4ce18cc985aaebf39abbb39400317fa19f4f710971e73bcab5805ec9a9"} err="failed to get container status \"68e2db4ce18cc985aaebf39abbb39400317fa19f4f710971e73bcab5805ec9a9\": rpc error: code = NotFound desc = could not find container \"68e2db4ce18cc985aaebf39abbb39400317fa19f4f710971e73bcab5805ec9a9\": container with ID starting with 68e2db4ce18cc985aaebf39abbb39400317fa19f4f710971e73bcab5805ec9a9 not found: ID does not exist" Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.588263 4815 scope.go:117] "RemoveContainer" containerID="33f27c7f2ff401aacb9573d7e3729ac7d2a6b530d98267f5546daf4606c05b3a" Feb 25 14:25:41 crc kubenswrapper[4815]: E0225 14:25:41.588770 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33f27c7f2ff401aacb9573d7e3729ac7d2a6b530d98267f5546daf4606c05b3a\": container with ID starting with 33f27c7f2ff401aacb9573d7e3729ac7d2a6b530d98267f5546daf4606c05b3a not found: ID does not exist" containerID="33f27c7f2ff401aacb9573d7e3729ac7d2a6b530d98267f5546daf4606c05b3a" Feb 25 14:25:41 crc kubenswrapper[4815]: I0225 14:25:41.588800 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33f27c7f2ff401aacb9573d7e3729ac7d2a6b530d98267f5546daf4606c05b3a"} err="failed to get container status \"33f27c7f2ff401aacb9573d7e3729ac7d2a6b530d98267f5546daf4606c05b3a\": rpc error: code = NotFound desc = could not find container \"33f27c7f2ff401aacb9573d7e3729ac7d2a6b530d98267f5546daf4606c05b3a\": container with ID starting with 33f27c7f2ff401aacb9573d7e3729ac7d2a6b530d98267f5546daf4606c05b3a not found: ID does not exist" Feb 25 14:25:42 crc kubenswrapper[4815]: I0225 14:25:42.996442 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84b81897-b136-4ce2-9c5a-6f3bae225e48" path="/var/lib/kubelet/pods/84b81897-b136-4ce2-9c5a-6f3bae225e48/volumes" Feb 25 14:25:45 crc kubenswrapper[4815]: I0225 14:25:45.936608 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:25:45 crc kubenswrapper[4815]: E0225 14:25:45.937386 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frc75_openshift-machine-config-operator(a593fa00-bfb6-44ec-bdbd-441921ae56c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-frc75" podUID="a593fa00-bfb6-44ec-bdbd-441921ae56c5" Feb 25 14:25:58 crc kubenswrapper[4815]: I0225 14:25:58.936666 4815 scope.go:117] "RemoveContainer" containerID="8a9b20c5e97c874bb4c6ca184cc85e2bdc471b3ebd7e12d9f194a0f963558107" Feb 25 14:25:59 crc kubenswrapper[4815]: I0225 14:25:59.686021 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frc75" event={"ID":"a593fa00-bfb6-44ec-bdbd-441921ae56c5","Type":"ContainerStarted","Data":"ab3e833ce735ac77959f242852cf18bb08b5d2e4daa5efd1dbf5895958ac5942"} Feb 25 14:26:00 crc kubenswrapper[4815]: I0225 14:26:00.147070 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533826-ln5zh"] Feb 25 14:26:00 crc kubenswrapper[4815]: E0225 14:26:00.147906 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84b81897-b136-4ce2-9c5a-6f3bae225e48" containerName="extract-content" Feb 25 14:26:00 crc kubenswrapper[4815]: I0225 14:26:00.147922 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="84b81897-b136-4ce2-9c5a-6f3bae225e48" containerName="extract-content" Feb 25 14:26:00 crc kubenswrapper[4815]: E0225 14:26:00.147975 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84b81897-b136-4ce2-9c5a-6f3bae225e48" containerName="extract-utilities" Feb 25 14:26:00 crc kubenswrapper[4815]: I0225 14:26:00.147987 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="84b81897-b136-4ce2-9c5a-6f3bae225e48" containerName="extract-utilities" Feb 25 14:26:00 crc kubenswrapper[4815]: E0225 14:26:00.148003 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84b81897-b136-4ce2-9c5a-6f3bae225e48" containerName="registry-server" Feb 25 14:26:00 crc kubenswrapper[4815]: I0225 14:26:00.148011 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="84b81897-b136-4ce2-9c5a-6f3bae225e48" containerName="registry-server" Feb 25 14:26:00 crc kubenswrapper[4815]: I0225 14:26:00.148272 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="84b81897-b136-4ce2-9c5a-6f3bae225e48" containerName="registry-server" Feb 25 14:26:00 crc kubenswrapper[4815]: I0225 14:26:00.149068 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533826-ln5zh" Feb 25 14:26:00 crc kubenswrapper[4815]: I0225 14:26:00.152251 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 14:26:00 crc kubenswrapper[4815]: I0225 14:26:00.155393 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533826-ln5zh"] Feb 25 14:26:00 crc kubenswrapper[4815]: I0225 14:26:00.156595 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 14:26:00 crc kubenswrapper[4815]: I0225 14:26:00.156777 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2wlcn" Feb 25 14:26:00 crc kubenswrapper[4815]: I0225 14:26:00.289966 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6f8k\" (UniqueName: \"kubernetes.io/projected/577bf20e-6101-4f88-9018-393249290ea9-kube-api-access-f6f8k\") pod \"auto-csr-approver-29533826-ln5zh\" (UID: \"577bf20e-6101-4f88-9018-393249290ea9\") " pod="openshift-infra/auto-csr-approver-29533826-ln5zh" Feb 25 14:26:00 crc kubenswrapper[4815]: I0225 14:26:00.392467 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6f8k\" (UniqueName: \"kubernetes.io/projected/577bf20e-6101-4f88-9018-393249290ea9-kube-api-access-f6f8k\") pod \"auto-csr-approver-29533826-ln5zh\" (UID: \"577bf20e-6101-4f88-9018-393249290ea9\") " pod="openshift-infra/auto-csr-approver-29533826-ln5zh" Feb 25 14:26:00 crc kubenswrapper[4815]: I0225 14:26:00.410960 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6f8k\" (UniqueName: \"kubernetes.io/projected/577bf20e-6101-4f88-9018-393249290ea9-kube-api-access-f6f8k\") pod \"auto-csr-approver-29533826-ln5zh\" (UID: \"577bf20e-6101-4f88-9018-393249290ea9\") " pod="openshift-infra/auto-csr-approver-29533826-ln5zh" Feb 25 14:26:00 crc kubenswrapper[4815]: I0225 14:26:00.483516 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533826-ln5zh" Feb 25 14:26:00 crc kubenswrapper[4815]: I0225 14:26:00.917597 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533826-ln5zh"] Feb 25 14:26:01 crc kubenswrapper[4815]: I0225 14:26:01.730268 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533826-ln5zh" event={"ID":"577bf20e-6101-4f88-9018-393249290ea9","Type":"ContainerStarted","Data":"a12f8a7fa82806c5d28f9260a5edd5c664fbd55e282db2ee534b0240b0c4e8be"} Feb 25 14:26:02 crc kubenswrapper[4815]: I0225 14:26:02.740346 4815 generic.go:334] "Generic (PLEG): container finished" podID="577bf20e-6101-4f88-9018-393249290ea9" containerID="f45dab2b8ef2af28e7939543cae4d800158edaa7a1180fa3941725d4e79c5138" exitCode=0 Feb 25 14:26:02 crc kubenswrapper[4815]: I0225 14:26:02.740494 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533826-ln5zh" event={"ID":"577bf20e-6101-4f88-9018-393249290ea9","Type":"ContainerDied","Data":"f45dab2b8ef2af28e7939543cae4d800158edaa7a1180fa3941725d4e79c5138"} Feb 25 14:26:04 crc kubenswrapper[4815]: I0225 14:26:04.104886 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533826-ln5zh" Feb 25 14:26:04 crc kubenswrapper[4815]: I0225 14:26:04.293008 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6f8k\" (UniqueName: \"kubernetes.io/projected/577bf20e-6101-4f88-9018-393249290ea9-kube-api-access-f6f8k\") pod \"577bf20e-6101-4f88-9018-393249290ea9\" (UID: \"577bf20e-6101-4f88-9018-393249290ea9\") " Feb 25 14:26:04 crc kubenswrapper[4815]: I0225 14:26:04.299015 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/577bf20e-6101-4f88-9018-393249290ea9-kube-api-access-f6f8k" (OuterVolumeSpecName: "kube-api-access-f6f8k") pod "577bf20e-6101-4f88-9018-393249290ea9" (UID: "577bf20e-6101-4f88-9018-393249290ea9"). InnerVolumeSpecName "kube-api-access-f6f8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 14:26:04 crc kubenswrapper[4815]: I0225 14:26:04.395576 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6f8k\" (UniqueName: \"kubernetes.io/projected/577bf20e-6101-4f88-9018-393249290ea9-kube-api-access-f6f8k\") on node \"crc\" DevicePath \"\"" Feb 25 14:26:04 crc kubenswrapper[4815]: I0225 14:26:04.766811 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533826-ln5zh" event={"ID":"577bf20e-6101-4f88-9018-393249290ea9","Type":"ContainerDied","Data":"a12f8a7fa82806c5d28f9260a5edd5c664fbd55e282db2ee534b0240b0c4e8be"} Feb 25 14:26:04 crc kubenswrapper[4815]: I0225 14:26:04.766884 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a12f8a7fa82806c5d28f9260a5edd5c664fbd55e282db2ee534b0240b0c4e8be" Feb 25 14:26:04 crc kubenswrapper[4815]: I0225 14:26:04.766897 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533826-ln5zh" Feb 25 14:26:05 crc kubenswrapper[4815]: I0225 14:26:05.198741 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533820-7bkpp"] Feb 25 14:26:05 crc kubenswrapper[4815]: I0225 14:26:05.212296 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533820-7bkpp"] Feb 25 14:26:06 crc kubenswrapper[4815]: I0225 14:26:06.946736 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbddfcbe-0ae8-4d91-a8d8-7d8314849143" path="/var/lib/kubelet/pods/fbddfcbe-0ae8-4d91-a8d8-7d8314849143/volumes" Feb 25 14:26:50 crc kubenswrapper[4815]: I0225 14:26:50.566639 4815 scope.go:117] "RemoveContainer" containerID="a00b2f27fcab78b9a064f917480a60e47bdd039de2649e0f7721bbbe4ff69466"